An Introduction to Stochastic Attribute-Value Grammars
by Rob Malouf, Miles Osborne
Publisher: ESSLLI 2001
Number of pages: 159
This one-week course will provide an introduction to the maximum entropy principle and the construction of maximum entropy models for natural language processing. Through a combination of lectures and, as local computing facilities permit, hands-on lab exercises, students will investigate the implementation of maximum entropy models for attribute-value grammars, including such topics as ambiguity identification, feature selection, and model training and evaluation.
Home page url
Download or read it online for free here:
by Doug Arnold, at al. - Blackwell Pub
This introductory book looks at all aspects of Machine Translation: covering questions of what it is like to use a modern Machine Translation system, through questions about how it is done, to questions of evaluating systems, and more.
by Dan Jurafsky, James H. Martin - Stanford University
This text takes an empirical approach to the subject, based on applying statistical and machine-learning algorithms to large corporations. The authors describe a unified vision of speech and language processing. Emphasis is on practical applications.
by Steven Bird, Ewan Klein, Edward Loper - O'Reilly Media
This book offers a highly accessible introduction to natural language processing, the field that supports a variety of language technologies. With it, you'll learn how to write Python programs that work with large collections of unstructured text.
by Roger Levy - University of California, San Diego
A book on the topic of using probabilistic models in scientific work on language ranging from experimental data analysis to corpus work to cognitive modeling. The intended audience is graduate students in linguistics, psychology and computer science.