Releases: scikit-learn-contrib/metric-learn
0.7.0
What's Changed
- Update paper ref after JMLR acceptance by @bellet in #301
- [MRG] Fix test for components_from_metric and add tests for _check_sdp_from_eigen by @wdevazelhes in #303
- Fix CI build - flake8 E231 by @terrytangyuan in #304
- [MRG+2] Update repo to work with both new and old scikit-learn by @wdevazelhes in #313
- [MRG+1] add github issues buttons by @wdevazelhes in #316
- [MRG] Fix urls in issue buttons by @bellet in #319
- Fix 4 tests by @mvargas33 in #321
- Addressing LFDA sign indeterminacy by @mvargas33 in #326
- YML file: CI Migration to Github Actions by @mvargas33 in #327
- Tidy up files to reflect the migration to GitHub Actions by @terrytangyuan in #328
- SCML iris test refactor [Minor] by @mvargas33 in #332
- Change min sklearn version to 0.21.3 by @mvargas33 in #334
- Disallow 0 on Triplets predictions by @mvargas33 in #331
- [Minor] Change SCML supervised docs by @mvargas33 in #337
score_pairs
refactor by @mvargas33 in #333- Fix 7 sources of warnings in the tests by @mvargas33 in #339
- [Tests][Warnings] Cut all warnings from SCML using a minimal solution by @mvargas33 in #341
- Threshold must be a real number by @mvargas33 in #322
- [Minor] GUI for matplotlib to run examples by @mvargas33 in #323
- [MRG][DOC] Fixes almost all warnings in the docs by @mvargas33 in #338
- Update SCML docstring by @nikosmichas in #347
- SCML: Raise ValueError if n_features larger than n_triplets by @nikosmichas in #350
- Rename variables, proposed by issue #257 by @mvargas33 in #324
- [MRG] Fix inaccurate SCML doctring for 'triplet_diffs' option by @bellet in #351
New Contributors
- @mvargas33 made their first contribution in #321
- @nikosmichas made their first contribution in #347
Full Changelog: v0.6.2...v0.7.0
0.6.2
This release uniformizes well version numbers that were mistaken in the previous release.
0.6.1
This release explicitly requires python>=3.6
and scikit-learn>=0.20.3
to install it.
0.6.0
This release features various fixes and improvements, as well as a new triplet-based algorithm, SCML (see http://researchers.lille.inria.fr/abellet/papers/aaai14.pdf), and an associated Triplets API. Triplets-based metric learning algorithms are used in settings where we have an "anchor" sample that we want to be closer with a "positive" sample than with a "negative" sample. Consistently with related packages like scikit-learn, we have also dropped support for Python 2 and Python 3.5.
New algorithms
- Add Sparse Compositional Metric Learning (SCML) (#278)
General updates on the package
- Drop support for python 2 and python 3.5 (#291)
- Add the Triplets API (#279)
- Solve issues in the documentation (#265, #266, #271, #274, #280)
- Allow installation from conda (#283)
- Fix covariance initialization when matrix is not invertible (#277)
- Add more robusts checks that an estimator is fitted (#267)
Improvements to existing algorithms
0.5.0
This is a major release in which the API (in particular for weakly-supervised algorithms) was largely refurbished in order to make it more unified and largely compatible with scikit-learn. Note that for this reason, you might encounter a significant amount of DeprecationWarning
and ChangedBehaviourWarning
. These warnings will disappear in version 0.6.0. The changes are summarized below:
-
All algorithms:
- Uniformize initialization for all algorithms: all algorithms that have a 'prior' or an 'init' as a parameter, can now choose it in a unified way, between (more) different choices ('identity', 'random', etc...) (#195 )
- Rename
num_dims
ton_components
for algorithms that have such a parameter. (#193) metric()
method has been renamed intoget_mahalanobis_matrix
(#152)- You can now use the function
score_pairs
to score a bunch of pair of points (return the distance between them), orget_metric
to get a metric function that can be plugged into scikit-learn estimators like any scipy distance.
-
Weakly supervised algorithms
- major API changes (#139, #217, #220, #197, #168) allowing greater compatibility with scikit-learn routines:
- in order to
fit
weakly supervised algorithms, users now have to provide 3d arrays of tuples (and possibly an array of labelsy
). For pairs learners, instead ofX
and[a, b, c, d]
as before, we should have an arraypairs
such thatpairs[i] = X[a[k], b[k]]
ify[i] == 1
orX[c[k], d[k]] if y[i] != 1
, wherek
is some integer (you can obtain such a representation by stacking horizontallya
andb
, thenc
andd
, stacking these vertically, and taking X[this array of indices]). For quadruplets learners, one should have the same form of input, instead that there is no need fory
, and that the 3d array will be an array of 4-uples instead of 2-uples. The two first elements of each quadruplet are the ones that we want to be more similar to each other than the last two. - Alternatively, a "preprocessor" can be used, if users instead want to give tuples of indices and not tuples of plain points, for less redundant manipulation of data. Custom preprocessor can be easily written for advanced use (e.g., to load and encode images from file paths).
- You can also use
predict
on a given pair or quadruplet, i.e. predict whether the pair is similar or not, or in the case of quadruplets, whether a given new quadruplet is in the right ordering or not - For pairs, this prediction depends on a threshold that can be set with
set_threshold
and calibrated on some data withcalibrate_threshold
. - For pairs, a default
score
is defined, which is the AUC (Area under the ROC Curve). For quadruplets, the defaultscore
is the accuracy (proportion of quadruplets given in the right order). - All of the above allows the algorithms to be compatible with scikit-learn for cross-validation, grid-search etc...
- For more information about these changes, see the new documentation
- in order to
- major API changes (#139, #217, #220, #197, #168) allowing greater compatibility with scikit-learn routines:
-
Supervised algorithms
- deprecation of
num_labeled
parameter (#119): - ITML_supervised bounds must now be set in init and not fit anymore (#163)
- deprecation of
use_pca
in LMNN (#231). - the random seed for generating constraints has now to be put at initialization rather than fit time (#224).
- removed preprocessing the data for RCA (#194).
- removed shogun dependency for LMNN (#216).
- deprecation of
-
Improved documentation:
-
Bug fixes:
- scikit-learn's fix scikit-learn/scikit-learn#13276 fixed SDML when the matrix to reconstruct is PSD, and the use of skggm fixed it in cases where the matrix is not PSD but we can still converge. The use of skggm is now recommended (i.e. we recommend to install skggm to use SDML).
- For all the algorithms that had a parameter
num_dims
(renamed ton_components
, see above), it will now be checked to be between 1 andn_features
, withn_features
the number of dimensions of the input space - LMNN did not update impostors at each iteration, which could result in problematic cases. Impostors are now recomputed at each iteration, which solves these problems (#228).
- The pseudo-inverse is now used in
Covariance
instead of the plain inverse, which allows to makeCovariance
work even in the case where the covariance matrix is not invertible (e.g. if the data lies on a space of smaller dimension).(#206) - There was an error in #101 that caused LMNN to return a wrong gradient (one dot product with
L
was missing). This has been fixed in #201.
0.4.0
- Two newly introduced algorithms:
- MLKR (Metric Learning for Kernel Regression)
- MMC (Mahalanobis Metric for Clustering)
- Improved documentation and examples
- Performance improvements
- Minor bug fixes
0.3.0
Constraints are now managed with a unified interface (metric_learn.Constraints
), which makes it easy to generate various input formats from (possibly) partial label information.
0.2.1
All classes inheriting from BaseMetricLearner
now support sklearn-style get_params
and set_params
.
0.2.0
We now support Python 3 alongside Python 2 in the same codebase.
0.1.1
This minor release adds two new methods:
- Local Fisher Discriminant Analysis (LFDA)
- Relative Components Analysis (RCA)
The performance of the non-Shogun LMNN implementation has also been improved, and it should now consume less memory.
This release also includes the new Sphinx documentation and improved docstrings for many of the classes and methods,