Document Type
Conference Proceeding
Publication Date
7-2008
Abstract
In real-world machine learning problems, it is very common that part of the input feature vector is incomplete: either not available, missing, or corrupted. In this paper, we present a boosting approach that integrates features with incomplete information and those with complete information to form a strong classifier. By introducing hidden variables to model missing information, we form loss functions that combine fully labeled data with partially labeled data to effectively learn normalized and unnormalized models. The primal problems of the proposed optimization problems with these loss functions are provided to show their close relationship and the motivations behind them. We use auxiliary functions to bound the change of the loss functions and derive explicit parameter update rules for the learning algorithms. We demonstrate encouraging results on two real-world problems — visual object recognition in computer vision and named entity recognition in natural language processing — to show the effectiveness of the proposed boosting approach.
Repository Citation
Haffari, G.,
Wang, Y.,
Wang, S.,
Mori, G.,
& Jiao, F.
(2008). Boosting with Incomplete Information. Proceedings of the 25th International Conference on Machine Learning, 368-375.
https://corescholar.libraries.wright.edu/knoesis/361
Included in
Bioinformatics Commons, Communication Technology and New Media Commons, Databases and Information Systems Commons, OS and Networks Commons, Science and Technology Studies Commons
Comments
Presented at the 25th International Conference on Machine Learning (ICML), Helsinki, Finland, July 5-9, 2008.