Doctoral thesis
OA Policy
English

Exploiting non-linear probabilistic models in natural language parsing and reranking

ContributorsTitov, Ivan
Defense date2008-01-28
Abstract

The thesis considers non-linear probabilistic models for natural language parsing, and it primarily focuses on the class of models which do not impose strict constraints on the structure of statistical dependencies. The main contribution is the demonstration that such models are appropriate for natural language parsing tasks and provide advantages over the use of standard 'linear' methods. We demonstrate this, first, by showing that though exact inference is intractable for the studied class of models, there exist accurate and tractable approximations. Second, we show that using non-linear representations results in powerful feature induction methods simplifying construction of parsers for new problems and domains, and leading to the state-of-the-art performance. Also, we demonstrate that the latent space induced by the model can be exploited in discriminative rerankers, and that this results in the significant improvement both in the standard parsing settings and in domain adaptation.

Keywords
  • Natural language parsing
  • Syntax
  • Dependency parsing
  • Constituent parsing
  • Natural language learning
  • Bayes risk minimization
  • Dynamic Bayesian networks
  • Variational approximations
Citation (ISO format)
TITOV, Ivan. Exploiting non-linear probabilistic models in natural language parsing and reranking. Doctoral Thesis, 2008. doi: 10.13097/archive-ouverte/unige:8495
Main files (1)
Thesis
accessLevelPublic
Identifiers
729views
336downloads

Technical informations

Creation06/29/2010 2:14:00 PM
First validation06/29/2010 2:14:00 PM
Update time03/14/2023 3:49:32 PM
Status update03/14/2023 3:49:32 PM
Last indexation10/29/2024 3:42:18 PM
All rights reserved by Archive ouverte UNIGE and the University of GenevaunigeBlack