2006 GraphBasedSemiSupervApproachForIE

From GM-RKB
Jump to navigation Jump to search

Subject Headings: Relation Recognition from Text Algorithm, ACE Benchmark Task, Semi-Supervised Leaerning Algorithm, Text Graph.

Notes

Cited By

Quotes

Abstract

Classification techniques deploy supervised labeled instances to train classifiers for various classification problems. However labeled instances are limited, expensive, and time consuming to obtain, due to the need of experienced human annotators. Meanwhile large amount of unlabeled data is usually easy to obtain. Semi-supervised learning addresses the problem of utilizing unlabeled data along with supervised labeled data, to build better classifiers. In this paper we introduce a semi-supervised approach based on mutual reinforcement in graphs to obtain more labeled data to enhance the classifier accuracy. The approach has been used to supplement a maximum entropy model for semi-supervised training of the ACE Relation Detection and Characterization (RDC) task. ACE RDC is considered a hard task in information extraction due to lack of large amounts of training data and inconsistencies in the available data. The proposed approach provides 10% relative improvement over the state of the art supervised baseline system.

Results and Discussion

We train several models like the one described in section 5.2 on different training data sets. In all experiments, we use both the LDC ACE training data and the labeled unsupervised data induced with the graph based approach we propose. We use the ACE evaluation procedure and ACE test corpus, provided by LDC, to evaluate all models.

We incrementally added labeled unsupervised data to the training data to determine the amount of data after which degradation in the system performance occurs. We sought this degradation point separately for each relation type. Figure 4 shows the effect of adding labeled unsupervised data on the ACE value for each relation separately. We notice from figure 4 and table 1 that relations with a small number of training instances had a higher gain in performance compared to relations with a large number of training instances. This implies that the proposed approach achieves significant improvement when the number of labeled training instances is small but representative.

References

ACE. (2004). The NIST ACE evaluation website. http://www.nist.gov/speech/tests/ace/ Avrim Blum, and Tom M. Mitchell. (1998). Combining Labeled and Unlabeled data with Co-training. Proceedings of the 11th Annual Conference on Computational Learning Theory. Avrim Blum and Shuchi Chawla. (2001). Learning From Labeled and Unlabeled Data Using Graph Mincuts. Proceedings of International Conference on Machine Learning (ICML). Avrim Blum, John D. Lafferty, Mugizi Rwebangira, and Rajashekar Reddy. (2004). Semi-supervised Learning Using Randomized Mincuts. Proceedings of the International Conference on Machine Learning (ICML). Stijn van Dongen. (2000). A Cluster Algorithm for Graphs. Technical Report INS-R0010, National Research Institute for Mathematics and Computer Science in the Netherlands. Stijn van Dongen. (2000). Graph Clustering by Flow Simulation. PhD thesis, University of Utrecht Radu Florian, Hany Hassan, Hongyan Jing, Nanda Kambhatla, Xiaqiang Luo, Nicolas Nicolov, and Salim Roukos. (2004). A Statistical Model for multilingual entity detection and tracking. Proceedings of the Human Language Technologies Conference (HLT-NAACL’04). Dayne Freitag, and Nicholas Kushmerick. 2000. Boosted wrapper induction. The 14th European Conference on Artificial Intelligence Workshop on Machine Learning for Information Extraction Taher Haveliwala. (2002). Topic-sensitive PageRank. Proceedings of the 11th International World Wide Web Conference Thorsten Joachims. (2003). Transductive Learning via Spectral Graph Partitioning. Proceedings of the International Conference on Machine Learning (ICML). John Kleinberg. (1998). Authoritative Sources in a Hyperlinked Environment. Proceedings of the. 9th ACM-SIAM Symposium on Discrete Algorithms. Nanda Kambhatla. (2004). Combining Lexical, Syntactic, and Semantic Features with Maximum Entropy Models for Information Extraction. Proceedings of the 42nd Annual Meeting of the Association for Computational Linguistics Ted Pedersen, Siddharth Patwardhan, and Jason Michelizzi, 2004, WordNet::Similarity - Measuring the Relatedness of Concepts. Proceedings of Fifth Annual Meeting of the North American Chapter of the Association for Computational Linguistics (NAACL- 2004) Scott White, and Padhraic Smyth. (2003). Algorithms for Discoveing Relative Importance in Graphs. Proceedings of Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Zhibiao Wu, and Martha Palmer. (1994). Verb semantics and lexical selection. Proceedings of the 32nd Annual Meeting of the Association for Computational Linguistics. Xiaojin Zhu, Zoubin Ghahramani, and John D. Lafferty. 2003. Semi-supervised Learning using Gaussian Fields and Harmonic Functions. Proceedings of the 20th International Conference on Machine Learning.


,

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2006 GraphBasedSemiSupervApproachForIEHany Hassan
Ahmed Hassan
Sara Noeman
Graph Based Semi-Supervised Approach for Information ExtractionProceedings of the HLT/NAACL-2006 Workshop on Graph-based methods for NLPhttp://www.computing.dcu.ie/~hhasan/semi-hlt06.pdf2006