Overview of the Research Field of Indefinite Learning
Motivation
Kernel methods like support vector machine (SVM), kernel principal component analysis (kPCA) or gaussian processes are popular tools in Machine Learning. The backbone of these approaches is the famous kernel trick, i.e., that a possibly infinite dimensional feature space is accessible via an inner product – the kernel function. The content object, all methods work with, is the kernel matrix, which must be positive-semi-definite (psd), so that a Hilbert space is induced. The field of indefinite learning emerges from the issue of indefinite kernels (kernel matrix or kernel function), which violate the assumptions of the kernel theory, by inducing a Krein space instead of a Hilbert space. Justifications for existence of such matrices can be given in the context of bioinformatics, local learning or consequences of approximation.
Main questions in this field are:
- What is the source of indefiniteness?
- Does it contain relevant information for the problem at hand?
- How to adopt the established methods so that they can handle these (non-metric/non-psd) matrices?
- How to optimize in the domain of non-convex loss-functions?
- How to correct the eigenspectrum of the indefinite matrix to be psd - effective and damage-free
Open Issues are often:
- providing a reasonable out-of-sample extension, so that the model can be used for unseen data
- improving on the approximation errors with efficient sampling strategies
- adopting the already established methods, in particular their optimization problem and most important, finding efficient algorithms in the context of large-scale learning
Relevant Researchers
Alphabetic (incomplete) list of researchers, who contributed to the field of indefinite learning. Research groups with co-authors are aligned.
Difeng Cai, Emory University
Gaelle Loosli, PobRun
Cheng Soon Ong, CSIRO
Ronny Luss, IBM
Alexandre D’aspremont, École normale supérieure - PSL
Katharina Morik, TU Dortmund
Dino Oglic, AstraZeneca
Thomas Gärtner, TU Wien
Elzbieta Pekalska, University of Manchester
Robert P.W. Duin, Delft University of Technology\ Bernard Haasdonk, Uni Stuttgart
Frank-Michael Schleif, UAS Würzburg
Peter Tino, University of Birmingham\ Andrej Gisbrecht, Aalto University\ Maximilian Münch, UAS Würzburg
Johan Suykens, KU Leuven
Siamak Mehrkanoon, Maastricht Univ.\ Fanghui Liu, Swiss Federal Institute of Technology in Lausanne\ Xiaolin Huang, Shanghai Jiao Tong Univ.
Recent Publications
Publications obtained via Google-Scholar (keywords: reproducing kernel krein space, indefinite nyström approximation, non-psd kernel, indefinite learning, indefinite SVM) since 2017 group as follows:
Large Scale Learning
Schleif, Frank-Michael, Andrej Gisbrecht, and Peter Tino. "Supervised low rank indefinite kernel approximation using minimum enclosing balls." Neurocomputing 318 (2018): 213-226.
Oglic, Dino, and Thomas Gärtner. "Scalable learning in reproducing kernel krein spaces." International Conference on Machine Learning. PMLR, 2019.
Andrej Gisbrecht, Frank-Michael Schleif. "Metric and non-metric proximity transformations at linear costs." Neurocomputing 167: 643-657 (2015)
Liu, Fanghui, et al. "Fast learning in reproducing kernel krein spaces via signed measures." International Conference on Artificial Intelligence and Statistics. PMLR, 2021.
Cai, Difeng, James Nagy, and Yuanzhe Xi. "Fast and stable deterministic approximation of general symmetric kernel matrices in high dimensions." arXiv preprint arXiv:2102.05215 (2021).
Heilig, Simon, Maximilian Münch and Frank-Michael Schleif. „Memory Efficient Kernel Approximation for Non-Stationary and Indefinite Kernels.” 2022 International Joint Conference on Neural Networks (IJCNN).
Krein Space Methods
Schleif, Frank-Michael, and Peter Tino. "Indefinite core vector machine." Pattern Recognition 71 (2017): 187-195.
Huang, Xiaolin, et al. "Indefinite kernels in least squares support vector machines and principal component analysis." Applied and Computational Harmonic Analysis 43.1 (2017): 162-172.
Huang, Xiaolin, et al. "Classification With Truncated L1 Distance Kernel." IEEE transactions on neural networks and learning systems 29.5 (2017): 2025-2030.
Mehrkanoon, Siamak, Xiaolin Huang, and Johan AK Suykens. "Indefinite kernel spectral learning." Pattern Recognition 78 (2018): 144-153.
Schleif, Frank-Michael, Christoph Raab, and Peter Tino. "Sparsification of indefinite learning models." Joint IAPR International Workshops on Statistical Techniques in Pattern Recognition (SPR) and Structural and Syntactic Pattern Recognition (SSPR). Springer, Cham, 2018.
Liu, Fanghui, et al. "Indefinite kernel logistic regression with concave-inexact-convex procedure." IEEE Transactions on Neural Networks and Learning Systems 30.3 (2018): 765-776.
Proxy Learning, Spectral Transformation, Embedding Methods
Liu, Fanghui, et al. "Robust kernel approximation for classification." International Conference on Neural Information Processing. Springer, Cham, 2017.
Münch, Maximilian, Christoph Raab, and Frank-Michael Schleif. "Encoding of Indefinite Proximity Data: A Structure Preserving Perspective." International Conference on Pattern Recognition Applications and Methods. Springer, Cham, 2020.
Münch, Maximilian, et al. "Scalable embedding of multiple perspectives for indefinite life-science data analysis." 2021 IEEE Symposium Series on Computational Intelligence (SSCI). IEEE, 2021.
Available Open-Source Software
- https://gitlab.ub.uni-bielefeld.de/bpaassen/python-edit-distances
- https://www.techfak.uni-bielefeld.de/~fschleif/software.xhtml#
- https://github.com/simonheilig/indefinite-meka
- http://37steps.com/distools/
- https://doglic.bitbucket.io/software/
- https://www.di.ens.fr/~aspremon/IndefiniteSVM.html
- http://gaelle.loosli.fr/code/KSVM.zip