Multi-View learning algorithms aim to learn from different representational views. For example, a movie can be represented by three views. The sequence of images, the audio, and the subtitles. Instead of concatenating the features of every view and training a single model, the Multi-View Stacking algorithm[1] builds independent (and possibly of different types) models for each view. These models are called first-level-learners. Then, the class and score predictions of the first-level-learners are used as features to train another model called the meta-learner. This approach is based on the Stacked Generalization method proposed by Wolpert D. H.[2].
The multiviewstacking
package provides the following functionalities:
- Train Multi-View Stacking classifiers.
- Supports arbitrary number of views. The limit is your computer's memory.
- Use any scikit-learn classifier as first-level-learner and meta-learner.
- Use any custom model as long as they implement the
fit()
,predict()
, andpredict_proba()
methods. - Combine different types of first-level-learners.
- Comes with a pre-loaded dataset with two views for testing.
- Python 3.8+
- pandas
- numpy
- scikit-learn >= 1.2.2
You can install the multiviewstacking
package with:
pip install multiviewstacking
This quick start example shows you how to train a multi-view model. For more detailed tutorials, check the jupyter notebooks in the /examples directory.
import numpy as np
from multiviewstacking import load_example_data
from multiviewstacking import MultiViewStacking
from sklearn.ensemble import RandomForestClassifier
# Load the built-in example dataset.
(xtrain,ytrain,xtest,ytest,ind1,ind2,l) = load_example_data()
The built-in dataset contains features for two views (audio, accelerometer) for activity recognition.
The load_example_data()
method returns a tuple with the train and test sets. It also returns the column indices for the two views and a LabelEnconder to convert the classes from integers back to strings.
# Define two first-level-learners and the meta-learner.
# All of them are Random Forests but they can be any other model.
m_v1 = RandomForestClassifier(n_estimators=50, random_state=123)
m_v2 = RandomForestClassifier(n_estimators=50, random_state=123)
m_meta = RandomForestClassifier(n_estimators=50, random_state=123)
# Train the model.
model = MultiViewStacking(views_indices = [ind1, ind2],
first_level_learners = [m_v1, m_v2],
meta_learner = m_meta)
The view_indices
parameter is a list of lists. Each list specifies the column indices of the train set for each view.
In this case ind1
stores the indices of the audio features and ind2
contains the indices of the accelerometer features.
Th first_level_learners
parameter is a list of scikit-learn models or any other custom models. The meta-learnr
specifies the model to be used as the meta-learner.
# Train the model.
model.fit(xtrain, ytrain)
# Make predictions on the test set.
preds = model.predict(xtest)
# Compuet the accuracy.
np.sum(ytest == preds) / len(ytest)
To cite this package use:
Enrique Garcia-Ceja (2024). multiviewstacking: A python implementation of the Multi-View Stacking algorithm.
Python package https://github.com/enriquegit/multiviewstacking
BibTex entry for LaTeX:
@Manual{MVS,
title = {multiviewstacking: A python implementation of the Multi-View Stacking algorithm},
author = {Enrique Garcia-Ceja},
year = {2024},
note = {Python package},
url = {https://github.com/enriquegit/multiviewstacking}
}
[1] Garcia-Ceja, Enrique, et al. "Multi-view stacking for activity recognition with sound and accelerometer data." Information Fusion 40 (2018): 45-56.
[2] Wolpert, D. H. (1992). Stacked generalization. Neural networks, 5(2), 241-259.