skip to main content
10.3115/1119176.1119206dlproceedingsArticle/Chapter ViewAbstractPublication PagesconllConference Proceedingsconference-collections
Article
Free Access

Early results for named entity recognition with conditional random fields, feature induction and web-enhanced lexicons

Published:31 May 2003Publication History

ABSTRACT

Models for many natural language tasks benefit from the flexibility to use overlapping, non-independent features. For example, the need for labeled data can be drastically reduced by taking advantage of domain knowledge in the form of word lists, part-of-speech tags, character n-grams, and capitalization patterns. While it is difficult to capture such inter-dependent features with a generative probabilistic model, conditionally-trained models, such as conditional maximum entropy models, handle them well. There has been significant work with such models for greedy sequence modeling in NLP (Ratnaparkhi, 1996; Borthwick et al., 1998).

References

  1. A. Borthwick, J. Sterling, E. Agichtein, and R. Grishman. 1998. Exploiting diverse knowledge sources via maximum entropy in named entity recognition. In Proceedings of the Sixth Workshop on Very Large Corpora, Association for Computational Linguistics.Google ScholarGoogle Scholar
  2. M. Collins and Y. Singer. 1999. Unsupervised models for named entity classification. In Proceedings of the Joint SIGDAT Conference on Empirical Methods in Natural Language Processing and Very Large Corpora.Google ScholarGoogle Scholar
  3. Stephen Della Pietra, Vincent J. Della Pietra, and John D. Lafferty. 1997. Inducing Features of Random Fields. IEEE Transactions on Pattern Analysis and Machine Intelligence, 19(4):380--393. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. Rosie Jones, Andrew McCallum, Kamal Nigam, and Ellen Riloff. 1999. Bootstrapping for Text Learning Tasks. In IJCAI-99 Workshop on Text Mining: Foundations, Techniques and Applications.Google ScholarGoogle Scholar
  5. John Lafferty, Andrew McCallum, and Fernando Pereira. 2001. Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data. In Proc. ICML. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. Robert Malouf. 2002. A comparison of algorithms for maximum entropy parameter estimation. In Sixth Workshop on Computational Language Learning (CoNLL-2002). Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Andrew McCallum and Fang-Fang Feng. 2003. Chinese Word Segmentation with Conditional Random Fields and Integrated Domain Knowledge. In Unpublished Manuscript.Google ScholarGoogle Scholar
  8. Andrew McCallum. 2003. Efficiently Inducing Features of Conditional Random Fields. In Nineteenth Conference on Uncertainty in Artificial Intelligence (UAI03). (Submitted). Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. Adwait Ratnaparkhi. 1996. A Maximum Entropy Model for Part-of-Speech Tagging. In Eric Brill and Kenneth Church, editors, Proceedings of the Conference on Empirical Methods in Natural Language Processing, pages 133--142. Association for Computational Linguistics.Google ScholarGoogle Scholar
  10. Fei Sha and Fernando Pereira. 2003. Shallow Parsing with Conditional Random Fields. In Proceedings of Human Language Technology, NAACL. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image DL Hosted proceedings
    CONLL '03: Proceedings of the seventh conference on Natural language learning at HLT-NAACL 2003 - Volume 4
    May 2003
    213 pages

    Publisher

    Association for Computational Linguistics

    United States

    Publication History

    • Published: 31 May 2003

    Qualifiers

    • Article

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader