An Introduction to Stochastic Attribute-Value Grammars
by Rob Malouf, Miles Osborne
Publisher: ESSLLI 2001
Number of pages: 159
This one-week course will provide an introduction to the maximum entropy principle and the construction of maximum entropy models for natural language processing. Through a combination of lectures and, as local computing facilities permit, hands-on lab exercises, students will investigate the implementation of maximum entropy models for attribute-value grammars, including such topics as ambiguity identification, feature selection, and model training and evaluation.
Home page url
Download or read it online for free here:
by Steven Bird, Ewan Klein, Edward Loper - O'Reilly Media
This book offers a highly accessible introduction to natural language processing, the field that supports a variety of language technologies. With it, you'll learn how to write Python programs that work with large collections of unstructured text.
by Paul Vogt - Language Science Press
This book presents a series of experiments in which two robots try to solve the symbol grounding problem. The experiments are based on the language game paradigm, and involve real mobile robots that are able to develop a grounded lexicon ...
by Jon Barwise, John Etchemendy - Center for the Study of Language
The book covers the boolean connectives, formal proof techniques, quantifiers, basic set theory, induction, proofs of soundness and completeness for propositional and predicate logic, and an accessible sketch of Godel's first incompleteness theorem.
by Shuly Wintner - ESSLLI
This text is a mild introduction to Formal Language Theory for students with little or no background in formal systems. The motivation is Natural Language Processing, and the presentation is geared towards NLP applications, with extensive examples.