Multiviewstacking
A python implementation of the Multi-View Stacking algorithm
Install / Use
/learn @enriquegit/MultiviewstackingREADME
multiviewstacking: a python implementation of the Multi-View Stacking algorithm <img src="https://github.com/enriquegit/multiviewstacking/blob/main/img/logo60-50.png?raw=true" align="right" width="300px " alt=""/>
In machine learning, Multi-View learning algorithms aim to learn from different representational views. For example, a movie can be represented by three views. The sequence of images, the audio, and the subtitles. Instead of concatenating the features of every view and training a single model, the Multi-View Stacking algorithm[1] builds independent (and possibly of different types) models for each view. These models are called first-level-learners. Then, the class and score predictions of the first-level-learners are used as features to train another model called the meta-learner. This approach is based on the Stacked Generalization method proposed by Wolpert D. H.[2].
The multiviewstacking package provides the following functionalities:
- Train Multi-View Stacking classifiers.
- Supports arbitrary number of views. The limit is your computer's memory.
- Use any scikit-learn classifier as first-level-learner and meta-learner.
- Use any custom model as long as they implement the
fit(),predict(), andpredict_proba()methods. - Combine different types of first-level-learners.
- Comes with a pre-loaded dataset with two views for testing.
:clipboard: Requirements
- Python >= 3.11.0
- pandas >= 2.0.0
- numpy >= 1.26.4
- scikit-learn >= 1.5.2
:wrench: Installation
You can install the multiviewstacking package with:
pip install multiviewstacking
:rocket: Quick start example
This quick start example shows you how to train a multi-view model. For more detailed tutorials, check the jupyter notebooks in the /examples directory.
import numpy as np
from multiviewstacking import load_example_data
from multiviewstacking import MultiViewStacking
from sklearn.ensemble import RandomForestClassifier
# Load the built-in example dataset.
(xtrain,ytrain,xtest,ytest,ind1,ind2,l) = load_example_data()
The built-in dataset contains features for two views (audio, accelerometer) for activity recognition.
The load_example_data() method returns a tuple with the train and test sets. It also returns the column indices for the two views and a LabelEnconder to convert the classes from integers back to strings.
# Define two first-level-learners and the meta-learner.
# All of them are Random Forests but they can be any other model.
m_v1 = RandomForestClassifier(n_estimators=50, random_state=123)
m_v2 = RandomForestClassifier(n_estimators=50, random_state=123)
m_meta = RandomForestClassifier(n_estimators=50, random_state=123)
# Train the model.
model = MultiViewStacking(views_indices = [ind1, ind2],
first_level_learners = [m_v1, m_v2],
meta_learner = m_meta)
The view_indices parameter is a list of lists. Each list specifies the column indices of the train set for each view.
In this case ind1 stores the indices of the audio features and ind2 contains the indices of the accelerometer features.
Th first_level_learners parameter is a list of scikit-learn models or any other custom models. The meta-learner specifies the model to be used as the meta-learner.
# Train the model.
model.fit(xtrain, ytrain)
# Make predictions on the test set.
preds = model.predict(xtest)
# Compute the accuracy.
np.sum(ytest == preds) / len(ytest)
Datasets for Multi-View Learning
Below there is a list of other datasets suitable for Multi-View learning models. Note that you may need to pre-process them and select the appropriate views.
- Multiple Features: Handwritten digits represented by 6 features categories (views). https://doi.org/10.24432/C5HC70
- HAR70+: Human Activity Recognition data collected with two 3-axial accelerometers. Each accelerometer can serve as a view. https://doi.org/10.24432/C5CW3D
- UNSW-NB15: Contains examples of network attacks for cybersecurity tasks. The features can be categorized into different views such as flow features, content features, time features, etc. https://research.unsw.edu.au/projects/unsw-nb15-dataset
Citation
To cite this package use:
Enrique Garcia-Ceja (2024). multiviewstacking: A python implementation of the Multi-View Stacking algorithm.
Python package https://github.com/enriquegit/multiviewstacking
BibTex entry for LaTeX:
@Manual{MVS,
title = {multiviewstacking: A python implementation of the Multi-View Stacking algorithm},
author = {Enrique Garcia-Ceja},
year = {2024},
note = {Python package},
url = {https://github.com/enriquegit/multiviewstacking}
}
References
[1] Garcia-Ceja, Enrique, et al. "Multi-view stacking for activity recognition with sound and accelerometer data." Information Fusion 40 (2018): 45-56.
[2] Wolpert, D. H. (1992). Stacked generalization. Neural networks, 5(2), 241-259.
