Skip to content

Latest commit

 

History

History
41 lines (26 loc) · 1.34 KB

README.md

File metadata and controls

41 lines (26 loc) · 1.34 KB

NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks

This is the official repository for the paper NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks. This repository contains the code for the experiments in the paper.

Installation

From PyPI

We provide a simple way to install the package from PyPI. You can install the package by running the following command:

pip install neuzip

Note that we only upload the source distribution to PyPI. You need to have NVCC correctly installed on your system to compile the package.

From source

You can also install the package from source, which is useful if you want to modify the code.

git clone https://github.com/BorealisAI/neuzip
cd neuzip
pip install -e .

Basic usage

Using neuzip for your PyTorch model is pretty easy. Here is a simple example:

model: torch.nn.Module = # your model
+ manager = neuzip.Manager()
+ model = manager.convert(model)

The compressed model can be used in the same way as the original model while consuming less memory.

Replicating experiments

You can replicate all the experiments in the paper by using the files in the examples/ directory. Each file corresponds to one or more experiments in the paper.