Conditional Maximum Entropy Model
Jump to navigation
Jump to search
See: Conditional Model, Maximum Entropy Classification Model, Maximum Entropy Principle, ME Model Parameter Estimation Algorithm.
References
2002
- (Malouf, 2002) ⇒ Robert Malouf. (2002). “A Comparison of Algorithms for Maximum Entropy Parameter Estimation.” In: Proceedings of the 6th conference on Natural language learning (COLING 2002). doi:10.3115/1118853.1118871
- ABSTRACT: Conditional maximum entropy (ME) models provide a general purpose machine learning technique which has been successfully applied to fields as diverse as computer vision and econometrics, and which is used for a wide variety of classification problems in natural language processing. However, the flexibility of ME models is not without cost. While parameter estimation for ME models is conceptually straightforward, in practice ME models for typical natural language tasks are very large, and may well contain many thousands of free parameters. In this paper, we consider a number of algorithms for estimating the parameters of ME models, including iterative scaling, gradient ascent, conjugate gradient, and variable metric methods. Surprisingly, the standardly used iterative scaling algorithms perform quite poorly in comparison to the others, and for all of the test problems, a limited-memory variable metric algorithm outperformed the other choices.