2006 MultilabelNeuralNetworksWithApplications

From GM-RKB
Jump to navigation Jump to search

Subject Headings: Multilabel Learning, Neural Network Algorithm, Backpropagation, Functional Genomics, Text Categorization

Notes

Cited By

Quotes

Abstract

In multilabel learning, each instance in the training set is associated with a set of labels and the task is to output a label set whose size is unknown a priori for each unseen instance. In this paper, this problem is addressed in the way that a neural network algorithm named BP-MLL, i.e., Backpropagation for Multilabel Learning, is proposed. It is derived from the popular Backpropagation algorithm through employing a novel error function capturing the characteristics of multilabel learning, i.e., the labels belonging to an instance should be ranked higher than those not belonging to that instance. Applications to two real-world multilabel learning problems, i.e., functional genomics and text categorization, show that the performance of BP-MLL is superior to that of some well-established multilabel learning algorithms.


References

  • [1] Christopher M. Bishop, Neural Networks for Pattern Recognition. New York: Oxford Univ. Press, 1995.
  • [2] M. R. Boutell, Jiebo Luo, X. Shen, and C.M. Brown, “Learning Multi-Label Scene Classification,” Pattern Recognition, vol. 37, no. 9, pp. 1757-1771, 2004.
  • [3] A. Clare, “Machine Learning and Data Mining for Yeast Functional Genomics,” PhD dissertation, Dept. of Computer Science, Univ. of Wales Aberystwyth, 2003.
  • [4] A. Clare and R.D. King, “Knowledge Discovery in Multi-Label Phenotype Data,” Lecture Notes in Computer Science, L.D. Raedt and A. Siebes, eds., vol. 2168, pp. 42-53. Berlin: Springer, 2001.
  • [5] F.D. Comité, R. Gilleron, and M. Tommasi, “Learning Multi-Label Alternating Decision Tree from Texts and Data,” Lecture Notes in Computer Science, P. Perner and A. Rosenfeld, eds., vol. 2734, pp. 35-49. Berlin: Springer, 2003.
  • [6] Arthur P. Dempster, N.M. Laird, and D.B. Rubin, “Maximum Likelihood from Incomplete Data via the EM Algorithm,” J. Royal Statistics Soc. B, vol. 39, no. 1, pp. 1-38, 1977.
  • [7] S.T. Dumais, J. Platt, David Heckerman, and M. Sahami, “Inductive Learning Algorithms and Representation for Text Categorization,” Proceedings of Seventh ACM Int'l Conference Information and Knowledge Management (CIKM '98), pp. 148-155, 1998.
  • [8] A. Elisseeff and J. Weston, “A Kernel Method for Multi-Labelled Classification,” Advances in Neural Information Processing Systems, Thomas G. Dietterich, S. Becker, and Zoubin Ghahramani, eds., vol. 14, pp. 681-687, 2002.
  • [9] Yoav Freund and L. Mason, “The Alternating Decision Tree Learning Algorithm,” Proceedings of 16th Int'l Conference Machine Learning (ICML '99), pp. 124-133, 1999.
  • [10] Yoav Freund and Robert E. Schapire, “A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting,” J. Computer and System Sciences, vol. 55, no. 1, pp. 119-139, 1997.
  • [11] S. Gao, W. Wu, C.-H. Lee, and T.-S. Chua, “A Maximal Figure-of-Merit Learning Approach to Text Categorization,” Proceedings of 26th ACM SIGIR Conference (SIGIR '03), pp. 174-181, 2003.
  • [12] S. Gao, W. Wu, C.-H. Lee, and T.-S. Chua, “A MFoM Learning Approach to Robust Multiclass Multi-Label Text Categorization,” Proceedings of 21st Int'l Conference Machine Learning (ICML '04), pp. 329-336, 2004.
  • [13] S. Haykin, Neural Networks: A Comprehensive Foundation, second ed. Englewood Cliffs, N.J.: Prentice-Hall, 1999.
  • [14] R. Jin and Zoubin Ghahramani, “Learning with Multiple Labels,” Advances in Neural Information Processing Systems, S. Becker, S. Thrun, and K. Obermayer, eds., vol. 15, pp. 897-904, 2003.
  • [15] Thorsten Joachims and F. Sebastiani, guest editors' introduction, J. Intelligent Information Systems, special issue on automated text categorization, vol. 18, nos. 2/3, pp. 103-105, Mar.-May 2002.
  • [16] H. Kazawa, T. Izumitani, H. Taira, and E. Maeda, “Maximal Margin Labeling for Multi-Topic Text Categorization,” Advances in Neural Information Processing Systems, L.K. Saul, Yair Weiss, and L. Bottou, eds., vol. 17, pp. 649-656, 2005.
  • [17] T. Kohonen, “An Introduction to Neural Computing,” Neural Networks, vol. 1, no. 1, pp. 3–16, 1988.
  • [18] Andrew McCallum, “Multi-Label Text Classification with a Mixture Model Trained by EM,” Proceedings of Working Notes Am. Assoc. Artificial Intelligence Workshop Text Learning (AAAI '99), 1999.
  • [19] W.S. McCulloch and W. Pitts, “A Logical Calculus of the Ideas Immanent in Nervous Activity,” Bull. Math. Biophysics, vol. 5, pp. 115-133, 1943.
  • [20] M. Minsky and S. Papert, Perceptrons. Cambridge, Mass.: MIT Press, 1969.
  • [21] P. Pavlidis, J. Weston, J. Cai, and W.N. Grundy, “Combining Microarray Expression Data and Phylogenetic Profiles to Learn Functional Categories Using Support Vector Machines,” Proceedings of Fifth Annual International Conference Computational Molecular Biology (RECOMB '01), pp. 242-248, 2001.
  • [22] J.R. Quinlan, Programs for Machine Learning. San Mateo, Calif.: Morgan Kaufmann, 1993.
  • [23] Frank Rosenblatt, Principles of Neurodynamics: Perceptrons and the Theory of Brain Mechanisms. Washington, D.C.: Spartan Books, 1962.
  • [24] D.E. Rumelhart, G.E. Hinton, and R.J. Williams, “Learning Internal Representations by Error Propagation,” Parallel Distributed Processing: Explorations in the Microstructure of Cognition, D.E. Rumelhart and J. L. McClelland, eds., vol. 1, pp. 318-362, 1986.
  • [25] Gerard M. Salton, Automatic Text Processing: The Transformation, Analysis, and Retrieval of Information by Computer. Reading, Pa.: Addison-Wesley, 1989.
  • [26] Gerard M. Salton, “Developments in Automatic Text Retrieval,” Science, vol. 253, pp. 974-980, 1991.
  • [27] Robert E. Schapire and Yoram Singer, “Improved Boosting Algorithms Using Confidence-Rated Predictions,” Proceedings of 11th Ann. Conference Computational Learning Theory (COLT '98), pp. 80-91, 1998.
  • [28] Robert E. Schapire and Yoram Singer, “BoosTexter: A Boosting-based System for Text Categorization,” Machine Learning, vol. 39, no. 2/3, pp. 135-168, 2000.
  • [29] F. Sebastiani, “Machine Learning in Automated Text Categorization,” ACM Computing Surveys, vol. 34, no. 1, pp. 1-47 2002.
  • [30] N. Ueda and K. Saito, “Parametric Mixture Models for Multi-Label Text,” Advances in Neural Information Processing Systems, S. Becker, S. Thrun, and K. Obermayer, eds., vol. 15, pp. 721-728, 2003.
  • [31] B. Widrow and M.E. Hoff, “Adaptive Switching Circuits,” IRE WESCON Convention Record, vol. 4, pp. 96-104, 1960.
  • [32] Y. Yang, “An Evaluation of Statistical Approaches to Text Categorization,” Information Retrieval, vol. 1, no. 1-2, pp. 69-90, 1999.
  • [33] Y. Yang and J.O. Pedersen, “A Comparative Study on Feature Selection in Text Categorization,” Proceedings of 14th Int'l Conference Machine Learning (ICML '97), pp. 412-420, 1997.
  • [34] Z.-H. Zhou, J. Wu, and W. Tang, “Ensembling Neural Networks: Many Could Be Better than All,” Artificial Intelligence, vol. 137, no. 1-2, pp. 239-263, 2002.

,

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2006 MultilabelNeuralNetworksWithApplicationsZhi-Hua Zhou
Min-Ling Zhang
Multilabel Neural Networks with Applications to Functional Genomics and Text CategorizationIEEE Transactions on Knowledge and Data Engineeringhttp://cs.nju.edu.cn/zhouzh/zhouzh.files/publication/tkde06a.pdf10.1109/TKDE.2006.1622006