Structured Penalties for Log-linear Language Models.

Authors
  • NELAKANTI Anil
  • ARCHAMBEAU Cedric
  • MAIRAL Julien
  • BACH Francis
  • BOUCHARD Guillaume
Publication date
2013
Publication type
Proceedings Article
Summary Language models can be formalized as loglinear regression models where the input features represent previously observed contexts up to a certain length m. The complexity of existing algorithms to learn the parameters by maximum likelihood scale linearly in nd, where n is the length of the training corpus and d is the number of observed features. We present a model that grows logarithmically in d, making it possible to efficiently leverage longer contexts. We account for the sequential structure of natural language using treestructured penalized objectives to avoid overfitting and achieve better generalization.
Topics of the publication
  • ...
  • No themes identified
Themes detected by scanR from retrieved publications. For more information, see https://scanr.enseignementsup-recherche.gouv.fr