Maksim Lapin (PhD Student)
MSc Maksim Lapin
- Address
- Max-Planck-Institut für Informatik
Saarland Informatics Campus
Campus - Location
- -
- Phone
- Fax
- Get email via email
Personal Information
Research Interests
- Computer Vision (image classification)
- Machine Learning (kernel methods)
Research Projects
- Loss Functions for Top-k Error: Analysis and Insights
- Top-k Multiclass SVM
- Efficient Output Kernel Learning for Multiple Tasks
- Scalable Multitask Representation Learning for Scene Classification
- Learning Using Privileged Information: SVM+ and Weighted SVM
Teaching
- Teaching Assistant, Machine Learning, Winter Semester 2014 (taught by Mario Fritz and Bjoern Andres)
- Teaching Assistant, Probabilistic Graphical Models and their Applications, Winter Semester 2013/2014 (taught by Bernt Schiele and Bjoern Andres)
- Teaching Assistant, Machine Learning, Winter Semester 2010/2011 (taught by Matthias Hein)
Education
- 2012–present, Ph.D. candidate in Computer Science, Max Planck Institute for Informatics
- 2012, M.Sc. in Computer Science, Saarland University
- 2006, Diploma in Mathematics, Belarusian State University
Personal Pages
Publications
2018
2017
Image Classification with Limited Training Data and Class Ambiguity
M. Lapin
PhD Thesis, Universität des Saarlandes, 2017
M. Lapin
PhD Thesis, Universität des Saarlandes, 2017
Abstract
Modern image classification methods are based on supervised learning algorithms that require labeled training data. However, only a limited amount of annotated data may be available in certain applications due to scarcity of the data itself or high costs associated with human annotation. Introduction of additional information and structural constraints can help improve the performance of a learning algorithm. In this thesis, we study the framework of learning using privileged information and demonstrate its relation to learning with instance weights. We also consider multitask feature learning and develop an efficient dual optimization scheme that is particularly well suited to problems with high dimensional image descriptors. Scaling annotation to a large number of image categories leads to the problem of class ambiguity where clear distinction between the classes is no longer possible. Many real world images are naturally multilabel yet the existing annotation might only contain a single label. In this thesis, we propose and analyze a number of loss functions that allow for a certain tolerance in top k predictions of a learner. Our results indicate consistent improvements over the standard loss functions that put more penalty on the first incorrect prediction compared to the proposed losses. All proposed learning methods are complemented with efficient optimization schemes that are based on stochastic dual coordinate ascent for convex problems and on gradient descent for nonconvex formulations.
2016
2015
Efficient Output Kernel Learning for Multiple Tasks
P. Jawanpuria, M. Lapin, M. Hein and B. Schiele
Advances in Neural Information Processing Systems 28 (NIPS 2015), 2015
P. Jawanpuria, M. Lapin, M. Hein and B. Schiele
Advances in Neural Information Processing Systems 28 (NIPS 2015), 2015
Top-k Multiclass SVM
M. Lapin, M. Hein and B. Schiele
Advances in Neural Information Processing Systems 28 (NIPS 2015), 2015
M. Lapin, M. Hein and B. Schiele
Advances in Neural Information Processing Systems 28 (NIPS 2015), 2015
2014
Scalable Multitask Representation Learning for Scene Classification
M. Lapin, B. Schiele and M. Hein
Scene Understanding Workshop (SUNw 2014), 2014
M. Lapin, B. Schiele and M. Hein
Scene Understanding Workshop (SUNw 2014), 2014