Skip to main content
Log in

MILDE: multiple instance learning by discriminative embedding

  • Regular Paper
  • Published:
Knowledge and Information Systems Aims and scope Submit manuscript

Abstract

While the objective of the standard supervised learning problem is to classify feature vectors, in the multiple instance learning problem, the objective is to classify bags, where each bag contains multiple feature vectors. This represents a generalization of the standard problem, and this generalization becomes necessary in many real applications such as drug activity prediction, content-based image retrieval, and others. While the existing paradigms are based on learning the discriminant information either at the instance level or at the bag level, we propose to incorporate both levels of information. This is done by defining a discriminative embedding of the original space based on the responses of cluster-adapted instance classifiers. Results clearly show the advantage of the proposed method over the state of the art, where we tested the performance through a variety of well-known databases that come from real problems, and we also included an analysis of the performance using synthetically generated data.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

Notes

  1. Furthermore, in practice, the repetition of elements occurs very rarely. This is due to the fact that the elements \(\vec {x}_j\) are almost always defined as vectors of real components, i.e., \(\vec {x}_j \in \mathbb R^d\) for \(j = 1,\ldots ,N.\)

  2. Given a multi-class problem, it can be reduced to several binary classification problems by means of common strategies such as one-against-all.

  3. In order to simplify the discussion, we consider only clusters of positive instances in this explanation. However, the same idea can be used in clusters of negative instances. In this case, a cluster is contaminated if not only contains negative instances but it also contains many positive ones.

  4. Regarding the negative instances, they were generated by using a constant number of Gaussian components, \(N_n=32\). We used a large number of components for the negative class in order to obtain realistic data. In this sense, note that the negative class is usually defined as the negation of the positive class, i.e., it groups all the classes of objects that are not positive, and thus forms an heterogeneous class.

References

  1. Dietterich TG, Lathrop RH, Lozano-Perez T (1997) Solving the multiple-instance problem with axis-parallel rectangles. Artif Intell 89:31–71

    Article  MATH  Google Scholar 

  2. Kotsiantis S, Kanellopoulos D (2008) Multi-instance learning for bankruptcy prediction. In: Proceedings of international conference on convergence and hybrid information technology, pp 1007–1012

  3. Chen Y, Bi J, Wang JZ (2006) MILES: multiple-instance learning via embedded instance selection. IEEE Trans Pattern Anal Mach Intell 28:1931–1947

    Article  Google Scholar 

  4. Reynolds DA, Quatieri TF, Dunn RB (2000) Speaker verification using adapted Gaussian mixture models. Dig Signal Process 10:19–41

    Article  Google Scholar 

  5. Oded M (1998) Learning from ambiguity. PhD thesis, MIT

  6. Zhang Q, Goldman SA (2001) EM-DD: an improved multiple-instance learning technique. In: Proceedings of neural information processing systems, pp 1073–1080

  7. Andrews S, Tsochantaridis I, Hofmann T (2003) Support vector machines for multiple-instance learning. In: Proceedings of neural information processing systems, pp 561–568

  8. Xu X, Frank E (2004) Logistic regression and boosting for labeled bags of instances. In: Proceedings of Pacific Asia conference on knowledge discovery and data mining, pp 272–281

  9. Gehler PV, Chapelle O (2007) Deterministic annealing for multiple-instance learning. In: Proceedings of international conference on artificial intelligence and statistics, pp 123–130

  10. Antic B, Ommer B (2013) Robust multiple-instance learning with superbags. Lecture notes in computer science, vol 7725, pp 242–255

  11. Scott S, Zhang J, Brown J (2005) On generalized multiple-instance learning. Int J Comput Intell Appl 5:21–35

    Article  Google Scholar 

  12. Zhou Z-H, Zhang M-L (2007) Solving multi-instance problems with classifier ensemble based on constructive clustering. Knowl Inf Syst 11:155–170

    Article  Google Scholar 

  13. Gärtner T, Flach PA, Kowalczyk A, Smola AJ (2002) Multi-instance kernels. In: Proceedings of international conference on machine learning, pp 179–186

  14. Foulds J, Frank E (2010) A review of multi-instance learning assumptions. Knowl Eng Rev 25:1–25

    Article  Google Scholar 

  15. Nowak E, Jurie F, Triggs B (2006) Sampling strategies for bag-of-features image classification. In: Proceedings of European conference on computer vision, pp 490–503

  16. Mikolajczyk K, Tuytelaars T, Schmid C, Zisserman A, Matas J, Schaffalitzky F, Kadir T, Gool LV (2005) A comparison of affine region detectors. Int J Comput Vis 65:43–72

    Article  Google Scholar 

  17. Maron O, Lozano-Pérez T (1998) A framework for multiple-instance learning. In: Proceedings of neural information processing systems, pp 570–576

  18. Bunescu R, Mooney R (2007) Multiple instance learning for sparse positive bags. In: Proceedings of international conference on machine learning, pp 105–112

  19. Zhou Z-H, Sun Y-Y, Li Y-F (2009), Multi-instance learning by treating instances as non i.i.d. samples. In: Proceedings of international conference on machine learning, pp 1249–1256

  20. Zhang M-L, Zhou Z-H (2009) Multi-instance clustering with applications to multi-instance prediction. Appl Intell 31:47–68

    Article  Google Scholar 

  21. Fan R-E, Chang K-W, Hsieh C-J, Wang X-R, Lin C-J (2008) Liblinear: a library for large linear classification. J Mach Learn Res 9:1871–1874

    MATH  Google Scholar 

  22. Hsu C-W, Chang C-C, Lin C-J (2003) A practical guide to support vector classification. Technical report, National Taiwan University

  23. Yang J (2005) Mill: a multi-instance learning library. http://www-2.cs.cmu.edu/juny/IR-lab/

Download references

Acknowledgments

We thank anonymous reviewers for their very useful comments and suggestions. This work was supported by the fellowship RYC-2008-03789 and the Spanish project TRA2011-29454-C03-01.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jaume Amores.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Amores, J. MILDE: multiple instance learning by discriminative embedding. Knowl Inf Syst 42, 381–407 (2015). https://doi.org/10.1007/s10115-013-0711-1

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10115-013-0711-1

Keywords

Navigation