skip to main content
10.3115/1073012.1073029dlproceedingsArticle/Chapter ViewAbstractPublication PagesaclConference Proceedingsconference-collections
Article
Free access

Immediate-head parsing for language models

Published: 06 July 2001 Publication History

Abstract

We present two language models based upon an "immediate-head" parser --- our name for a parser that conditions all events below a constituent c upon the head of c. While all of the most accurate statistical parsers are of the immediate-head variety, no previous grammatical language model uses this technology. The perplexity for both of these models significantly improve upon the trigram model base-line as well as the best previous grammar-based language model. For the better of our two models these improvements are 24% and 14% respectively. We also suggest that improvement of the underlying parser should significantly improve the model's perplexity and that even in the near term there is a lot of potential for improvement in immediate-head language models.

References

[1]
Bod, R. What is the minimal set of fragments that achieves maximal parse accuracy. In Proceedings of Association for Computational Linguistics 2001. 2001.
[2]
Charniak, E. Treebank grammars. In Proceedings of the Thirteenth National Conference on Artificial Intelligence. AAAI Press/MIT Press, Menlo Park, 1996, 1031--1036.
[3]
Charniak, E. A maximum-entropy-inspired parser. In Proceedings of the 2000 Conference of the North American Chapter of the Association for Computational Linguistics. ACL, New Brunswick NJ, 2000.
[4]
Chelba, C. and Jelinek, F. Exploiting syntactic structure for language modeling. In Proceedings for COLING-ACL 98. ACL, New Brunswick NJ, 1998, 225--231.
[5]
Chi, Z. and Geman, S. Estimation of probabilistic context-free grammars. Computational Linguistics 242 (1998), 299--306.
[6]
Collins, M. J. Three generative lexicalized models for statistical parsing. In Proceedings of the 35th Annual Meeting of the ACL. 1997, 16--23.
[7]
Collins, M. J. Head-Driven Statistical Models for Natural Language Parsing. University of Pennsylvania, Ph.D. Dissertation, 1999.
[8]
Collins, M. J. Discriminative reranking for natural language parsing. In Proceedings of the International Conference on Machine Learning (ICML 2000). 2000.
[9]
Goddeau, D. Using probabilistic shift-reduce parsing in speech recognition systems. In Proceedings of the 2nd International Conference on Spoken Language Processing. 1992, 321--324.
[10]
Goodman, J. Putting it all together: language model combination. In ICASSP-2000. 2000.
[11]
Lauer, M. Corpus statistics meet the noun compound: some empirical results. In Proceedings of the 33rd Annual Meeting of the Association for Computational Linguistics. 1995, 47--55.
[12]
Magerman, D. M. Statistical decision-tree models for parsing. In Proceedings of the 33rd Annual Meeting of the Association for Computational Linguistics. 1995, 276--283.
[13]
Marcus, M. P., Santorini, B. and Marcinkiewicz, M. A. Building a large annotated corpus of English: the Penn treebank. Computational Linguistics 19 (1993), 313--330.
[14]
Ratnaparkhi, A. Learning to parse natural language with maximum entropy models. Machine Learning 34 1/2/3 (1999), 151--176.
[15]
Roark, B. Probabilistic top-down parsing and language modeling. Computational Linguistics (forthcoming).
[16]
Stolcke, A. An efficient probabilistic context-free parsing algorithm that computes prefix probabilities. Computational Linguistics 21 (1995), 165--202.
[17]
Stolcke, A. and Segal, J. Precise ngram probabilities from stochastic context-free grammars. In Proceedings of the 32th Annual Meeting of the Association for Computational Linguistics. 1994, 74--79.

Cited By

View all

Recommendations

Comments

Information & Contributors

Information

Published In

cover image DL Hosted proceedings
ACL '01: Proceedings of the 39th Annual Meeting on Association for Computational Linguistics
July 2001
562 pages

Publisher

Association for Computational Linguistics

United States

Publication History

Published: 06 July 2001

Qualifiers

  • Article

Acceptance Rates

Overall Acceptance Rate 85 of 443 submissions, 19%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)84
  • Downloads (Last 6 weeks)5
Reflects downloads up to 26 Jan 2025

Other Metrics

Citations

Cited By

View all

View Options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Login options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media