skip to main content
10.3115/1119355.1119374dlproceedingsArticle/Chapter ViewAbstractPublication PagesemnlpConference Proceedingsconference-collections
Article
Free Access

Investigating loss functions and optimization methods for discriminative learning of label sequences

Published:11 July 2003Publication History

ABSTRACT

Discriminative models have been of interest in the NLP community in recent years. Previous research has shown that they are advantageous over generative models. In this paper, we investigate how different objective functions and optimization methods affect the performance of the classifiers in the discriminative learning framework. We focus on the sequence labelling problem, particularly POS tagging and NER tasks. Our experiments show that changing the objective function is not as effective as changing the features included in the model.

References

  1. Y. Altun, T. Hofmann, and M. Johnson. 2002. Discriminative learning for label sequences via boosting. In Proceedings of NIPS*15.Google ScholarGoogle Scholar
  2. Daniel M. Bikel, Richard L. Schwartz, and Ralph M. Weischedel. 1999. An algorithm that learns what's in a name. Machine Learning, 34(1--3):211--231. Google ScholarGoogle ScholarDigital LibraryDigital Library
  3. M. Collins. 2000. Discriminative reranking for natural language parsing. In Proceedings of ICML 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  4. M. Collins. 2002. Ranking algorithms for named-entity extraction: Boosting and the voted perceptron. In Proceedings of ACL'02. Google ScholarGoogle ScholarDigital LibraryDigital Library
  5. M. Johnson, S. Geman, S. Canon, Z. Chi, and S. Riezler. 1999. Estimators for stochastic unifi cation-based grammars. In Proceedings of ACL'99. Google ScholarGoogle ScholarDigital LibraryDigital Library
  6. S. Kakade, Y. W. Teh, and S. Roweis. 2002. An alternative objective function for Markovian fi elds. In Proceedings of ICML 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  7. Dan Klein and Christopher D. Manning. 2002. Conditional structure versus conditional estimation in nip models. In Proceedings of EMNLP 2002. Google ScholarGoogle ScholarDigital LibraryDigital Library
  8. J. Lafferty, A. McCallum, and F. Pereira. 2001. Conditional random fields: Probabilistic models for segmenting and labeling sequence data. In Proceedings of ICML2001. Google ScholarGoogle ScholarDigital LibraryDigital Library
  9. A. McCallum, D. Freitag, and F. Pereira. 2000. Maximum Entropy Markov Models for Information Extraction and Segmentation. In Proceedings of ICML 2000. Google ScholarGoogle ScholarDigital LibraryDigital Library
  10. T. Minka. 2001. Algorithms for maximum-likelihood logistic regression. Technical report, CMU, Department of Statistics, TR 758.Google ScholarGoogle Scholar
  11. V. Punyakanok and D. Roth. 2000. The use of classifiers in sequential inference. In Proceedings of NIPS*13.Google ScholarGoogle Scholar
  12. Adwait Ratnaparkhi. 1999. Learning to parse natural language with maximum entropy models. Machine Learning, 34(1-3):151--175. Google ScholarGoogle ScholarDigital LibraryDigital Library
  13. R. Schapire and Y. Singer. 1999. Improved boosting algorithms using confidence-rated predictions. Machine Learning, 37(3):297--336. Google ScholarGoogle ScholarDigital LibraryDigital Library
  14. Kristina Toutanova and Christopher Manning. 2000. Enriching the knowledge sources used in a maximum entropy pos tagger. In Proceedings of EMNLP 2000. Google ScholarGoogle ScholarDigital LibraryDigital Library

Recommendations

Comments

Login options

Check if you have access through your login credentials or your institution to get full access on this article.

Sign in
  • Published in

    cover image DL Hosted proceedings
    EMNLP '03: Proceedings of the 2003 conference on Empirical methods in natural language processing
    July 2003
    224 pages

    Publisher

    Association for Computational Linguistics

    United States

    Publication History

    • Published: 11 July 2003

    Qualifiers

    • Article

    Acceptance Rates

    Overall Acceptance Rate73of234submissions,31%

PDF Format

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader