2009 DistributionalRepresentationsfo

From GM-RKB
Jump to navigation Jump to search

Subject Headings: Supervised Sequence Labeling Algorithm

Notes

Cited By

Quotes

Abstract

Supervised sequence-labeling systems in natural language processing often suffer from data sparsity because they use word types as features in their prediction tasks. Consequently, they have difficulty estimating parameters for types which appear in the test set, but seldom (or never) appear in the training set. We demonstrate that distributional representations of word types, trained on unannotated text, can be used to improve performance on rare words. We incorporate aspects of these representations into the feature space of our sequence-labeling systems. In an experiment on a standard chunking dataset, our best technique improves a chunker from 0.76 F1 to 0.86 F1 on chunks beginning with rare words. On the same dataset, it improves our part-of-speech tagger from 74% to 80% accuracy on rare words. Furthermore, our system improves significantly over a baseline system when applied to text from a different domain, and it reduces the sample complexity of sequence labeling.

References

,

 AuthorvolumeDate ValuetitletypejournaltitleUrldoinoteyear
2009 DistributionalRepresentationsfoAlexander Yates
Fei Huang
Distributional Representations for Handling Sparsity in Supervised Sequence-labeling2009