- Welcome to the InstructLab CLI
- ❓ What is
ilab
- 📋 Requirements
- ✅ Getting started
- 💻 Creating new knowledge or skills and training the model
- 🎁 Contribute knowledge or compositional skills
- 📜 List and validate your new data
- 🚀 Generate a synthetic dataset
- 👩🏫 Training the model
- ✋ Before you begin training
- InstructLab training pipelines
- Train the model locally
- Train the model locally on an M-Series Mac or on Linux using the full pipeline
- Train the model locally on an M-Series Mac or on Linux using the simple pipeline
- Train the model locally with GPU acceleration
- Train the model in the cloud
- 📜 Test the newly trained model
- 🧪 Evaluate the newly trained model
- 🍴 Serve the newly trained model
- 📣 Chat with the new model (not optional this time)
- 🎁 Submit your new knowledge or skills
- 📬 Contributing
InstructLab 🐶 uses a novel synthetic data-based alignment tuning method for Large Language Models (LLMs.) The "lab" in InstructLab 🐶 stands for Large-Scale Alignment for ChatBots [1].
[1] Shivchander Sudalairaj*, Abhishek Bhandwaldar*, Aldo Pareja*, Kai Xu, David D. Cox, Akash Srivastava*. "LAB: Large-Scale Alignment for ChatBots", arXiv preprint arXiv: 2403.01081, 2024. (* denotes equal contributions)
ilab
is a Command-Line Interface (CLI) tool that allows you to perform the following actions:
- Download a pre-trained Large Language Model (LLM).
- Chat with the LLM.
To add new knowledge and skills to the pre-trained LLM, add information to the companion taxonomy repository.
After you have added knowledge and skills to the taxonomy, you can perform the following actions:
- Use
ilab
to generate new synthetic training data based on the changes in your localtaxonomy
repository. - Re-train the LLM with the new training data.
- Chat with the re-trained LLM to see the results.
graph TD;
download-->chat
chat[Chat with the LLM]-->add
add[Add new knowledge<br/>or skill to taxonomy]-->generate[generate new<br/>synthetic training data]
generate-->train
train[Re-train]-->|Chat with<br/>the re-trained LLM<br/>to see the results|chat
For an overview of the full workflow, see the workflow diagram.
Important
We have optimized InstructLab so that community members with commodity hardware can perform these steps. However, running InstructLab on a laptop will provide a low-fidelity approximation of synthetic data generation
(using the ilab data generate
command) and model instruction tuning (using the ilab model train
command, which uses QLoRA). To achieve higher quality, use more sophisticated hardware and configure InstructLab to use a
larger teacher model such as Mixtral.
-
🍎 Apple M1/M2/M3 Mac or 🐧 Linux system (tested on Fedora). We anticipate support for more operating systems in the future.
📋 When installing the
ilab
CLI on macOS, you may have to run thexcode-select --install
command, installing the required packages listed. -
C++ compiler
-
Python 3.10 or Python 3.11
📋 Python 3.12 is currently not supported. Some InstructLab dependencies don't work on Python 3.12, yet.
-
Minimum 250GB disk space. Approximately 500GB disk space is recommended for the entire InstructLab end-to-end process.
-
When installing on Fedora Linux, install C++, Python 3.10 or 3.11, and other necessary tools by running the following command:
sudo dnf install gcc gcc-c++ make git python3.11 python3.11-devel
Some Python version management tools that build Python (instead of using a pre-built binary) may not by default build libraries implemented in C, and CPython when installing a Python version. This can result in the following error when running the
ilab data generate
command:ModuleNotFoundError: No module named '_lzma'
. This can be resolved by building CPython during the Python installation with the--enable-framework
. For example forpyenv
on MacOS:PYTHON_CONFIGURE_OPTS="--enable-framework" pyenv install 3.x
. You may need to recreate your virtual environment after reinstalling Python.
Note
The following steps in this document use Python venv for virtual environments. However, if you use another tool such as pyenv or Conda Miniforge for managing Python environments on your machine continue to use that tool instead. Otherwise, you may have issues with packages that are installed but not found in your virtual environment.
-
There are a few ways you can locally install the
ilab
CLI. Select your preferred installation method from the following instructions. You can then installilab
and activate yourvenv
environment.⚠️ Thepython3
binary shown in the installation guides are the Python version that you installed in the above step. The command can also bepython3.11
orpython3.10
instead ofpython3
. You can check Python's version bypython3 -V
.⏳
pip install
may take some time, depending on your internet connection. In case the installation fails with errorunsupported instruction `vpdpbusd'
, append-C cmake.args="-DLLAMA_NATIVE=off"
topip install
command.See the GPU acceleration documentation for how to to enable hardware acceleration for interaction and training on AMD ROCm, Apple Metal Performance Shaders (MPS), and Nvidia CUDA.
-
Install on Apple metal with:
python<version> -m venv --upgrade-deps venv source venv/bin/activate pip cache remove llama_cpp_python pip install instructlab
📋 Make sure your system Python build is
Mach-O 64-bit executable arm64
by usingfile -b $(command -v python)
, or if your system is setup with pyenv by using thefile -b $(pyenv which python)
command.
-
Install on a standard Linux machine with:
python<version> -m venv --upgrade-deps venv source venv/bin/activate pip install instructlab
Additional Build Argument for Intel Macs If you have a Mac with an Intel CPU, you must add a prefix of
CMAKE_ARGS="-DLLAMA_METAL=off"
to thepip install
command to ensure that the build is done without Apple M-series GPU support.(venv) $ CMAKE_ARGS="-DLLAMA_METAL=off" pip install ...
-
Install on AMD ROCm with:
python3 -m venv --upgrade-deps venv source venv/bin/activate pip cache remove llama_cpp_python CMAKE_ARGS="-DLLAMA_HIPBLAS=on \ -DAMDGPU_TARGETS=all \ -DCMAKE_C_COMPILER=/opt/rocm/llvm/bin/clang \ -DCMAKE_CXX_COMPILER=/opt/rocm/llvm/bin/clang++ \ -DCMAKE_PREFIX_PATH=/opt/rocm \ -DLLAMA_NATIVE=off" \ pip install 'instructlab[rocm]' \ --extra-index-url https://download.pytorch.org/whl/rocm6.0
On Fedora 40+, use
-DCMAKE_C_COMPILER=clang-17
and-DCMAKE_CXX_COMPILER=clang++-17
.
-
For the best CUDA experience, installing vLLM is necessary to serve Safetensors format models.
python3 -m venv --upgrade-deps venv source venv/bin/activate pip cache remove llama_cpp_python CMAKE_ARGS="-DLLAMA_CUDA=on -DLLAMA_NATIVE=off" pip install 'instructlab[cuda]' pip install vllm@git+https://github.com/opendatahub-io/[email protected]
-
From your
venv
environment, verifyilab
is installed correctly, by running theilab
command.ilab
Example output of the
ilab
command(venv) $ ilab Usage: ilab [OPTIONS] COMMAND [ARGS]... CLI for interacting with InstructLab. If this is your first time running InstructLab, it's best to start with `ilab config init` to create the environment. Options: --config PATH Path to a configuration file. [default: /home/user/.config/instructlab/config.yaml] -v, --verbose Enable debug logging (repeat for even more verbosity) --version Show the version and exit. --help Show this message and exit. Commands: config Command Group for Interacting with the Config of InstructLab. data Command Group for Interacting with the Data generated by... model Command Group for Interacting with the Models in InstructLab. system Command group for all system-related command calls taxonomy Command Group for Interacting with the Taxonomy of InstructLab. Aliases: chat model chat generate data generate serve model serve train model train
Important
Every ilab
command needs to be run from within your Python virtual environment. You can enter the Python environment by running the source venv/bin/activate
command.
-
Optional: You can enable tab completion for the
ilab
command.Enable tab completion in
bash
with the following command:eval "$(_ILAB_COMPLETE=bash_source ilab)"
To have this enabled automatically every time you open a new shell, you can save the completion script and source it from
~/.bashrc
:_ILAB_COMPLETE=bash_source ilab > ~/.ilab-complete.bash echo ". ~/.ilab-complete.bash" >> ~/.bashrc
📋 To use Bash version 4.4 or higher on macOS (default is 3.2.57), ensure your login shell is set to the updated version. You can verify this with
echo $SHELL
. If you encounter the errorbash: complete: nosort: invalid option name
, check your terminal or configuration files (e.g., ~/.bash_profile, ~/.bashrc, ~/.profile) to see whether they are referencing the old version for login.Enable tab completion in
zsh
with the following command:eval "$(_ILAB_COMPLETE=zsh_source ilab)"
To have this enabled automatically every time you open a new shell, you can save the completion script and source it from
~/.zshrc
:_ILAB_COMPLETE=zsh_source ilab > ~/.ilab-complete.zsh echo ". ~/.ilab-complete.zsh" >> ~/.zshrc
Enable tab completion in
fish
with the following command:_ILAB_COMPLETE=fish_source ilab | source
To have this enabled automatically every time you open a new shell, you can save the completion script and source it from
~/.bashrc
:_ILAB_COMPLETE=fish_source ilab > ~/.config/fish/completions/ilab.fish
-
Initialize
ilab
by running the following command:ilab config init
-
When prompted, clone the
https://github.com/instructlab/taxonomy.git
repository into the current directory by typing enterOptional: If you want to point to an existing local clone of the
taxonomy
repository, you can pass the path interactively or alternatively with the--taxonomy-path
flag.ilab
will use the default configuration file unless otherwise specified. You can override this behavior with the--config
parameter for anyilab
command. -
When prompted, provide the path to your default model. Otherwise, the default of a quantized Granite model is used.
Example output of steps 1 - 3
Path to your model [/home/user/.cache/instructlab/models/granite-7b-lab-Q4_K_M.gguf]: <ENTER>
You can download this model with
ilab model download
command as well. -
When prompted, please choose a train profile. Train profiles are GPU specific profiles that enable accelerated training behavior. If you are on MacOS or a Linux machine without a dedicated GPU, please choose
No Profile (CPU, Apple Metal, AMD ROCm)
by hitting Enter. There are various flags you can utilize with individualilab
commands that allow you to utilize your GPU if applicable.Example output of selecting a training profile
---------------------------------------------------- Welcome to the InstructLab CLI This guide will help you to setup your environment ---------------------------------------------------- Please provide the following values to initiate the environment [press Enter for defaults]: Path to taxonomy repo [/home/user/.local/share/instructlab/taxonomy]: Path to your model [/home/user/.cache/instructlab/models/granite-7b-lab-Q4_K_M.gguf]: You can download this model with `ilab model download` command as well.
-
The InstructLab CLI auto-detects your hardware and select the exact system profile that matches your machine. System profiles populate the
config.yaml
file with the proper parameter values based on your detected GPU types and available vRAM.Example output of profile auto-detection
Generating config file and profiles: /home/user/.config/instructlab/config.yaml /home/user/.local/share/instructlab/internal/train_configuration/profiles We have detected the AMD CPU profile as an exact match for your system. -------------------------------------------- Initialization completed successfully! You're ready to start using `ilab`. Enjoy! --------------------------------------------
-
If there is not an exact match for your system, you can manually select a system profile when prompted. There are various flags you can utilize with individual
ilab
commands that allow you to utilize your GPU if applicable.Example output of selecting a system profile
Please choose a system profile to use. System profiles apply to all parts of the config file and set hardware specific defaults for each command. First, please select the hardware vendor your system falls into [1] APPLE [2] INTEL [3] AMD [4] NVIDIA Enter the number of your choice [0]: 1 You selected: APPLE Next, please select the specific hardware configuration that most closely matches your system. [0] No system profile [1] APPLE M1 ULTRA [2] APPLE M1 MAX [3] APPLE M2 MAX [4] APPLE M2 ULTRA [5] APPLE M2 PRO [6] APPLE M2 [7] APPLE M3 MAX [8] APPLE M3 PRO [9] APPLE M3 Enter the number of your choice [hit enter for hardware defaults] [0]: 8 You selected: /home/<user>/.local/share/instructlab/internal/system_profiles/apple/m3/m3_pro.yaml -------------------------------------------- Initialization completed successfully! You're ready to start using `ilab`. Enjoy! --------------------------------------------
The GPU profiles are listed by GPU type and number of GPUs present. If you happen to have a GPU configuration with a similar amount of vRAM as any of the above profiles, feel free to try them out!
After running ilab config init
your directories will look like the following on a Linux system:
Directory | Description |
---|---|
~/.cache/instructlab/models/ |
Contains all downloaded large language models, including the saved output of ones you generate with ilab. |
~/.local/share/instructlab/datasets/ |
Contains data output from the SDG phase, built on modifications to the taxonomy repository. |
~/.local/share/instructlab/taxonomy/ |
Contains the skill and knowledge data. |
~/.local/share/instructlab/checkpoints/ |
Contains the output of the training process. |
~/.config/instructlab/config.yaml |
Contains the config.yaml file |
You can view your config.yaml
file with the following command:
ilab config show
-
Run the
ilab model download
command to download a compact pre-trained version of thegranite-7b-lab-GGUF
,merlinite-7b-lab-GGUF
, andMistral-7B-Instruct-v0.2-GGUF
models (~4.4G each) from HuggingFace.ilab model download
Example output
Downloading model from Hugging Face: Model: instructlab/granite-7b-lab-GGUF@main Destination: /Users/<user>/.cache/instructlab/models Downloading model from Hugging Face: Model: instructlab/merlinite-7b-lab-GGUF@main Destination: /Users/<user>/.cache/instructlab/models Downloading model from Hugging Face: Model: TheBloke/Mistral-7B-Instruct-v0.2-GGUF@main Destination: /Users/<user>/.cache/instructlab/models TheBloke/Mistral-7B-Instruct-v0.2-GGUF requires a HF Token to be set. Please use '--hf-token' or 'export HF_TOKEN' to download all necessary models.
You may be prompted to use your Hugging Face token to download the
Mistral-7B-Instruct-v0.2-GGUF
model.ilab model download --hf-token <your-huggingface-token>
If you have issues connecting to Hugging Face, refer to the Hugging Face discussion forum for more details.
-
Specify a repository, model, and a Hugging Face token if necessary. More information about Hugging Face tokens can be found here
ilab model download --repository instructlab/granite-7b-lab-GGUF --filename granite-7b-lab-Q4_K_M.gguf --hf-token <your-huggingface-token>
-
Specify a repository, and a Hugging Face token if necessary. For example:
ilab model download --repository instructlab/granite-7b-lab --hf-token <your-hugginface-token>
These types of models are useful for GPU-enabled systems or anyone looking to serve a model using vLLM. InstructLab provides Safetensor versions of our Granite models on HuggingFace.
-
All downloaded models can be seen with
ilab model list
.ilab model list
Example output of
ilab model list
afterilab model download
(venv) $ ilab model list +-------------------------------------+---------------------+--------+ | Model Name | Last Modified | Size | +-------------------------------------+---------------------+--------+ | granite-7b-lab-Q4_K_M.gguf | 2024-08-01 15:05:48 | 4.1 GB | | merlinite-7b-lab-Q4_K_M.gguf | 2024-08-01 15:05:48 | 4.1 GB | | mistral-7b-instruct-v0.2.Q4_K_M.gguf| 2024-08-01 15:05:48 | 4.1 GB | +-------------------------------------+---------------------+--------+
-
Serve the model with the
ilab model serve
command:ilab model serve
Example output of a model that is served and ready
(venv) $ ilab model serve INFO 2024-03-02 02:21:11,352 lab.py:201 Using model 'models/ggml-granite-7b-lab-Q4_K_M.gguf' with -1 gpu-layers and 4096 max context size. Starting server process After application startup complete see http://127.0.0.1:8000/docs for API. Press CTRL+C to shut down the server.
📋 If multiple
ilab
clients try to connect to the same InstructLab server at the same time, the 1st will connect to the server while the others will start their own temporary server. This will require additional resources on the host machine.
-
You can serve a non-default model (e.g. Mixtral-8x7B-Instruct-v0.1) with the following example command:
ilab model serve --model-path ~/.cache/instructlab/models/mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf
-
You can serve a non-default Safetensors model (e.g. granite-7b-lab). This requires a GPU.
Verify vllm is installed:
pip show vllm
If it is not, please run:
pip install vllm@git+https://github.com/opendatahub-io/[email protected]
Then you can serve a Safetensors model:
ilab model serve --model-path ~/.cache/instructlab/models/instructlab/granite-7b-lab
-
Because you're serving the model in one terminal window, you will have to create a new terminal window and re-activate your Python virtual environment to run
ilab model chat
command:source venv/bin/activate ilab model chat
-
Chat with a non-default model (e.g. Mixtral-8x7B-Instruct-v0.1):
source venv/bin/activate ilab model chat --model ~/.cache/instructlab/models/mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf
[!TIP] The usage of
--model
necessitates that the existing server has that model. If not, you must exit the server.--model
inilab model chat
has the ability to start a server on your behalf with the specified model if one is not already running on the port.Before you start adding new skills and knowledge to your model, you can check its baseline performance by asking it a question such as
what is the capital of Canada?
. The model needs to be trained with the generated synthetic data to use the new skills or knowledgeExample output of chatbot
(venv) $ ilab model chat ╭────────────────────────────────────────────────────────────────────────────────────────────────────────────────── system ──────────────────────────────────────────────────────────────────────────────────────────────────────────────────╮ │ Welcome to InstructLab Chat w/ GGML-GRANITE-7B-lab-Q4_K_M (type /h for help) │ ╰────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────╯ >>> what is the capital of Canada? [S][default] ╭────────────────────────────────────────────────────────────────────────────────────────────────────── ggml-granite-7b-lab-Q4_K_M ───────────────────────────────────────────────────────────────────────────────────────────────────────╮ │ The capital city of Canada is Ottawa. It is located in the province of Ontario, on the southern banks of the Ottawa River in the eastern portion of southern Ontario. The city serves as the political center for Canada, as it is home to │ │ Parliament Hill, which houses the House of Commons, Senate, Supreme Court, and Cabinet of Canada. Ottawa has a rich history and cultural significance, making it an essential part of Canada's identity. │ ╰─────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────────── elapsed 12.008 seconds ─╯ >>> [S][default]
-
You can also adjust the temperature settings of the model with the
--greedy-mode
flag.ilab model chat --model ~/.cache/instructlab/models/mixtral-8x7b-instruct-v0.1.Q4_K_M.gguf --greedy-mode
-
To upgrade InstructLab to the latest version, use the following command:
pip install instructlab --upgrade
-
It is recommended to reinitialize your InstructLab environment after upgrading to the latest version.
ilab config init
- Contribute new knowledge or compositional skills to your local taxonomy repository.
Detailed contribution instructions can be found in the taxonomy repository.
Important
There is a limit to how much content can exist in the question/answer pairs for the model to process. Due to this, only add a maximum of around 2300 words to your question and answer seed example pairs in the qna.yaml
file.
-
You can use the
ilab taxonomy diff
command to ensureilab
is registering your new knowledge or skills and your contributions are properly formatted. This command displays any new or modified YAML files within your taxonomy tree. For example, the following is the expected result of a valid compositional skill contribution after adding a new skill calledfoo-lang
to the freeform writing subdirectory:(venv) $ ilab taxonomy diff compositional_skills/writing/freeform/foo-lang/qna.yaml Taxonomy in $HOME/.local/share/instructlab/taxonomy is valid :)
-
You can also validate your entire taxonomy by performing a diff against an empty base by using the
--taxonomy-base=empty
argument:(venv) $ ilab taxonomy diff --taxonomy-base=empty compositional_skills/general/tables/empty/qna.yaml compositional_skills/general/tables/editing/add_remove/qna.yaml ... Taxonomy in $HOME/.local/share/instructlab/taxonomy is valid :)
Before following these instructions, ensure the existing model you are adding skills or knowledge to is still running. Alternatively, ilab data generate
can start a server for you if you provide a fully qualified model path via --model
.
-
To generate a synthetic dataset based on your newly added knowledge or skill set in taxonomy repository, run the following command:
Default SDG
ilab data generate
With GPU acceleration
ilab data generate --pipeline full --gpus <NUM_OF_GPUS>
Without GPU acceleration:
ilab data generate --pipeline simple
Using a non-default model (e.g. Mixtral-8x7B-Instruct-v0.1) to generate data:
ilab data generate --model ~/.cache/instructlab/models/mistralai/mixtral-8x7b-instruct-v0.1 --pipeline full --gpus 4
⏳ This can take from 15 minutes to 1+ hours to complete, depending on your computing resources.
Example output of
ilab data generate
(venv) $ ilab data generate INFO 2024-07-30 19:57:44,093 numexpr.utils:161: NumExpr defaulting to 8 threads. INFO 2024-07-30 19:57:44,452 datasets:58: PyTorch version 2.3.1 available. Generating synthetic data using 'full' pipeline, 'mistral-7b-instruct-v0.2.Q4_K_M.gguf' model, './taxonomy' taxonomy, against http://localhost:8000/v1 server INFO 2024-07-30 19:57:45,084 instructlab.sdg:375: Synthesizing new instructions. If you aren't satisfied with the generated instructions, interrupt training (Ctrl-C) and try adjusting your YAML files. Adding more examples may help. INFO 2024-07-30 19:57:45,090 instructlab.sdg.pipeline:153: Running pipeline single-threaded INFO 2024-07-30 19:57:47,820 instructlab.sdg.llmblock:51: LLM server supports batched inputs: False INFO 2024-07-30 19:57:47,820 instructlab.sdg.pipeline:197: Running block: gen_skill_freeform INFO 2024-07-30 19:57:47,820 instructlab.sdg.pipeline:198: Dataset({ features: ['task_description', 'seed_question', 'seed_response'], num_rows: 5 }) INFO 2024-07-30 20:02:16,455 instructlab.sdg:411: Generated 1 samples ...
The synthetic data set will be two files in the newly created in the datasets directory:
~/.local/share/instructlab/datasets
. These files will be namedskills_train_msgs_*.jsonl
andknowledge_train_msgs_*.jsonl
. -
Verify the files have been created by running the
ls datasets
command. Note: you must be in yourXDG_DATA_HOME/instructlab
directory.(venv) $ ls datasets/ ├── node_datasets_2024-08-12T20_31_15 ├── node_datasets_2024-08-13T19_51_48 ├── knowledge_recipe_2024-08-12T20_31_15.yaml ├── knowledge_recipe_2024-08-13T19_51_48.yaml ├── knowledge_train_msgs_2024-08-12T20_31_15.jsonl ├── knowledge_train_msgs_2024-08-13T19_51_48.jsonl ├── skills_recipe_2024-08-12T20_31_15.yaml ├── skills_recipe_2024-08-13T19_51_48.yaml ├── skills_train_msgs_2024-08-12T20_31_15.jsonl ├── skills_train_msgs_2024-08-13T19_51_48.jsonl ├── messages_mixtral-8x7b-instruct-v0-1_2024-08-12T20_31_15.jsonl ├── messages_mixtral-8x7b-instruct-v0-1_2024-08-13T19_51_48.jsonl ├── test_mixtral-8x7b-instruct-v0-1_2024-08-12T20_13_21.jsonl ├── test_mixtral-8x7b-instruct-v0-1_2024-08-12T20_23_06.jsonl ├── test_mixtral-8x7b-instruct-v0-1_2024-08-12T20_31_15.jsonl ├── test_mixtral-8x7b-instruct-v0-1_2024-08-13T19_47_59.jsonl ├── test_mixtral-8x7b-instruct-v0-1_2024-08-13T19_51_48.jsonl ├── train_mixtral-8x7b-instruct-v0-1_2024-08-12T20_31_15.jsonl └── train_mixtral-8x7b-instruct-v0-1_2024-08-13T19_51_48.jsonl
Generating synthetic data on a different model: It is also possible to run the generate step against a different model via an
OpenAI-compatible API. For example, the one spawned by ilab model serve
or any remote or locally hosted LLM (e.g. via ollama
, LM Studio
, etc.). Run the following command:
ilab data generate --endpoint-url http://localhost:8000/v1
Generating synthetic data on the entire taxonomy repo: You can generate a synthetic dataset based on the entire contents of the taxonomy repo using the --taxonomy-base=empty
option:
ilab data generate --taxonomy-base=empty
There are many options for training the model with your synthetic data-enhanced dataset.
There are a few models you need to download before running the InstructLab end-to-end workflow locally.
-
Download the
granite-7b-lab
model for training:ilab model download --repository instructlab/granite-7b-lab
-
Download the
prometheus-8x7b-v2.0
for multi-phase training and benchmark evaluation. This model is not required forsimple
orfull
training.ilab model download --repository prometheus-eval/prometheus-8x7b-v2.0 --hf-token <your-huggingface-token>
ilab model train
has three pipelines: simple
, full
, and accelerated
. The default is full
.
simple
uses an SFT Trainer on Linux and MLX on MacOS. This type of training takes roughly an hour and produces the lowest fidelity model but should indicate if your data is being picked up by the training process.full
uses a custom training loop and data processing functions for the granite family of models. This loop is optimized for CPU and MPS functionality. Please use--pipeline=full
in combination with--device=cpu
(Linux) or--device=mps
(MacOS). You can also use--device=cpu
on a MacOS machine. However, MPS is optimized for better performance on these systems.accelerated
uses the instructlab-training library which supports GPU accelerated and distributed training. Thefull
loop and data processing functions are either pulled directly from or based off of the work in this library.
After running ilab model train
, the output locations depend on the chosen pipeline or strategy:
Pipeline/Strategy | Operating System | Output Location/Details |
---|---|---|
simple |
Linux | Model saved in models directory as ggml-model-f16.gguf . |
simple |
MacOS | Model saved in <model_name>-mlx-q directory. |
full |
Linux & MacOS | .bin and .gguf models saved in ~/.local/share/instructlab/checkpoints/hf_format directory. Two models in each sample_* directory: one quantized (Q4-M-K suffix) and one full precision. |
accelerated |
Linux | Models saved in ~/.local/share/instructlab/checkpoints . Can be evaluated with ilab model evaluate to choose the best one. |
lab-multiphase |
Linux | Phase 1 models saved in ~/.local/share/instructlab/phased/phase1/checkpoints (Knowledge training). Phase 2 models saved in ~/.local/share/instructlab/phased/phase2/checkpoints (Skills training). Evaluation is run for both phases to identify the best checkpoint. |
To limit training time, you can adjust the num_epoch
paramater in the config.yaml
file. The maximum number of epochs for running the InstructLab end-to-end workkflow is 10.
-
Train the model with your synthetic data with the
ilab model train
command:ilab model train
⏳ This step can potentially take several hours to complete depending on your computing resources. Please stop
ilab model chat
andilab model serve
first to free resources.
When running multi phase training evaluation is run on each phase, we will tell you which checkpoint in this folder performs the best.
-
To train the model locally on your M-Series Mac using our full pipeline and MPS or on your Linux laptop/desktop using CPU:
ilab model train --pipeline full --device mps --data-path <path-to-sdg-dataset>
ilab model train --pipeline full --device cpu --data-path <path-to-sdg-dataset>
Example command
ilab model train --pipeline full --device cpu --data-path ~/.local/share/instructlab/datasets/knowledge_train_msgs_2024-10-23T09_14_44.jsonl
⏳ This process will take a while to complete. If you run for ~8 epochs it will take several hours.
ilab model train
outputs a directory for each epoch that resembles the following structure:$ ls ~/.local/share/instructlab/checkpoints/hf_format/samples_0/ ├── added_tokens.json ├── config.json ├── pytorch_model.bin ├── pytorch_model.gguf ├── pytorch_model-Q4_K_M.gguf ├── special_tokens_map.json ├── tokenizer.json ├── tokenizer_config.json └── tokenizer.model
This entire folder can be served on a system that supports vLLM using the
.bin
model. However, on most laptops you can serve either the full precision gguf:pytorch_model.gguf
or the 4-bit-quantized one:pytorch_model-Q4_K_M.gguf
.
-
To train the model locally on your M-Series Mac using our simple pipeline and MLX or on your Linux laptop/desktop using an SFT Trainer:
ilab model train --pipeline simple
⏳ This process will take a little while to complete (time can vary based on hardware and output of
ilab data generate
but on the order of 5 to 15 minutes)On a Mac
ilab model train
outputs a brand-new model that is saved in the<model_name>-mlx-q
directory calledadapters.npz
(inNumpy
compressed array format). For example:(venv) $ ls instructlab-granite-7b-lab-mlx-q/ ├── adapters-010.npz ├── adapters-020.npz ├── adapters-030.npz ├── adapters-040.npz ├── adapters-050.npz ├── adapters-060.npz ├── adapters-070.npz ├── adapters-080.npz ├── adapters-090.npz ├── adapters-100.npz ├── adapters.npz ├── added_tokens.json ├── config.json ├── model.safetensors ├── special_tokens_map.json ├── tokenizer.json ├── tokenizer.model └── tokenizer_config.json
-
Training has support for GPU acceleration with Nvidia CUDA or AMD ROCm. Please see the GPU acceleration documentation for more details. At present, hardware acceleration requires a data center GPU or high-end consumer GPU with at least 18 GB free memory.
ilab model train --pipeline accelerated --device cuda --data-path <path-to-sdg-data>
Example command
ilab model train --pipeline full --device cpu --data-path ~/.local/share/instructlab/datasets/knowledge_train_msgs_2024-10-23T09_14_44.jsonl
This version of
ilab model train
outputs brand-new models that can be served in the~/.local/share/instructlab/checkpoints
directory. These models can be run throughilab model evaluate
to choose the best one.
-
ilab model train
supports multi-phase training. This results in the following workflow:- We train the model on knowledge
- Evaluate the trained model to find the best checkpoint
- We train the model on skills
- We evaluate the model to find the best overall checkpoint
ilab model train --strategy lab-multiphase --phased-phase1-data <knowledge train messages jsonl> --phased-phase2-data <skills train messages jsonl> -y
This command takes in two
.jsonl
files from yourdatasets
directory, one is the knowledge jsonl and the other is a skills jsonl. The-y
flag skips an interactive prompt asking the user if they are sure they want to run multi-phase training.⏳ This command may take 3 or more hours depending on the size of the data and number of training epochs you run.
Follow the instructions in Training.
⏳ Approximate amount of time taken on each platform:
- Google Colab: 5-10 minutes with a T4 GPU
- Kaggle: ~30 minutes with a P100 GPU.
After that's done, you can play with your model directly in the Google Colab or Kaggle notebook. Model trained on the cloud will be saved on the cloud. The model can also be downloaded and served locally.
-
Run the following command to test the model:
ilab model test
The output from the command will consist of a series of outputs from the model before and after training.
You can use the ilab model evaluate
command to evaluate the models you are training with several benchmarks. Currently, four benchmarks are supported.
Benchmark | Measures | Full Name | Description | Reference |
---|---|---|---|---|
MMLU | Knowledge | Massive Multitask Language Understanding | Tests a model against a standardized set of knowledge data and produces a score based on the model's performance | Measuring Massive Multitask Language Understanding |
MMLUBranch | Knowledge | N/A | Tests your knowledge contributions against a base model and produces a score based on the difference in performance | N/A |
MTBench | Skills | Multi-turn Benchmark | Tests a model's skill at applying its knowledge against a judge model and produces a score based on the model's performance | MT-Bench (Multi-turn Benchmark) |
MTBenchBranch | Skills | N/A | Tests your skill contributions against a judge model and produces a score based on the difference in performance | N/A |
Note
Evaluation must be used with local models (safetensors or GGUF format). Using models directly from Hugging Face without downloading them is unsupported.
GGUF models are not yet supported for mmlu and mmlu_branch evaluations
MTBench and MTBenchBranch use prometheus-8x7b-v2.0 as the judge model by
default. While you do not need to use this model as your judge, it is strongly recommended to do so if you have the necessary hardware
resources. You can download it via ilab model download
.
-
Example of running MMLU with a local safetensors model directory:
$ export ILAB_MODELS_DIR=$HOME/.local/share/instructlab/models $ ilab model evaluate --benchmark mmlu --model $ILAB_MODELS_DIR/instructlab/granite-7b-test ... # KNOWLEDGE EVALUATION REPORT ## MODEL (SCORE) /home/user/.cache/instructlab/models/instructlab/granite-7b-test (0.52/1.0) ### SCORES (0.0 to 1.0): mmlu_abstract_algebra - 0.31 mmlu_anatomy - 0.46 mmlu_astronomy - 0.52 mmlu_business_ethics - 0.55 mmlu_clinical_knowledge - 0.57 mmlu_college_biology - 0.56 mmlu_college_chemistry - 0.38 mmlu_college_computer_science - 0.46 mmlu_college_mathematics - 0.34 mmlu_college_medicine - 0.49 mmlu_college_physics - 0.27 mmlu_computer_security - 0.66 mmlu_conceptual_physics - 0.38 mmlu_econometrics - 0.39 mmlu_electrical_engineering - 0.48 mmlu_elementary_mathematics - 0.3 ...
The output of MMLU displays a much longer list of subjects.
-
Example of running MMLUBranch with a local safetensors model directory:
$ export ILAB_MODELS_DIR=$HOME/.local/share/instructlab/models $ ilab model evaluate --benchmark mmlu_branch --model $ILAB_MODELS_DIR/instructlab/granite-7b-test --base-model $ILAB_MODELS_DIR/instructlab/granite-7b-lab ... # KNOWLEDGE EVALUATION REPORT ## BASE MODEL (SCORE) /home/user/.cache/instructlab/models/instructlab/granite-7b-lab (0.74/1.0) ## MODEL (SCORE) /home/user/.cache/instructlab/models/instructlab/granite-7b-test (0.78/1.0) ### IMPROVEMENTS (0.0 to 1.0): 1. tonsils: 0.74 -> 0.78 (+0.04)
-
Example of running MTBench with a local safetensors model directory:
$ export ILAB_MODELS_DIR=$HOME/.local/share/instructlab/models $ ilab model evaluate --benchmark mt_bench --model $ILAB_MODELS_DIR/instructlab/granite-7b-test ... # SKILL EVALUATION REPORT ## MODEL (SCORE) /home/user/.cache/instructlab/models/instructlab/granite-7b-test (7.27/10.0) ### TURN ONE (0.0 to 10.0): 7.48 ### TURN TWO (0.0 to 10.0): 7.05
-
Below is an example of running MTBenchBranch with a local safetensors model directory:
$ export ILAB_MODELS_DIR=$HOME/.local/share/instructlab/models $ export ILAB_TAXONOMY_DIR=$HOME/.local/share/instructlab/taxonomy $ ilab model evaluate --benchmark mt_bench_branch \ --model $ILAB_MODELS_DIR/instructlab/granite-7b-test \ --base-model $ILAB_MODELS_DIR/instructlab/granite-7b-lab \ --taxonomy-path $ILAB_TAXONOMY_DIR \ --branch rc \ --base-branch main ... # SKILL EVALUATION REPORT ## BASE MODEL (SCORE) /home/user/.cache/instructlab/models/instructlab/granite-7b-lab (5.78/10.0) ## MODEL (SCORE) /home/user/.cache/instructlab/models/instructlab/granite-7b-test (6.00/10.0) ### IMPROVEMENTS (0.0 to 10.0): 1. foundational_skills/reasoning/linguistics_reasoning/object_identification/qna.yaml: 4.0 -> 6.67 (+2.67) 2. foundational_skills/reasoning/theory_of_mind/qna.yaml: 3.12 -> 4.0 (+0.88) 3. foundational_skills/reasoning/linguistics_reasoning/logical_sequence_of_words/qna.yaml: 9.33 -> 10.0 (+0.67) ### REGRESSIONS (0.0 to 10.0): 1. foundational_skills/reasoning/unconventional_reasoning/lower_score_wins/qna.yaml: 5.67 -> 4.0 (-1.67) 2. foundational_skills/reasoning/mathematical_reasoning/qna.yaml: 7.33 -> 6.0 (-1.33) 3. foundational_skills/reasoning/temporal_reasoning/qna.yaml: 5.67 -> 4.67 (-1.0) ### NO CHANGE (0.0 to 10.0): 1. foundational_skills/reasoning/linguistics_reasoning/odd_one_out/qna.yaml (9.33) 2. compositional_skills/grounded/linguistics/inclusion/qna.yaml (6.5)
If you have a server running, stop the server you have running by entering ctrl+c
keys in the terminal running the server.
🍎 Mac only: Before serving the newly trained model you must convert it to work with
the ilab
cli. The ilab model convert
command converts the new model into quantized GGUF format which is required by the server to host the model in the ilab model serve
command.
-
Convert the newly trained model by running the following command:
ilab model convert
-
Serve the newly trained model locally via
ilab model serve
command with the--model-path
argument to specify your new model:ilab model serve --model-path <new model path>
Which model should you select to serve? After running the
ilab model convert
command, some files and a directory are generated. The model you will want to serve ends with an extension of.gguf
and exists in a directory with the suffixtrained
. For example:instructlab-granite-7b-lab-trained/instructlab-granite-7b-lab-Q4_K_M.gguf
.
-
Try the fine-tuned model out live using the chat interface, and see if the results are better than the untrained version of the model with chat by running the following command:
ilab model chat -m <New model path>
If you are interested in optimizing the quality of the model's responses, please see
TROUBLESHOOTING.md
Of course, the final step is, if you've improved the model, to open a pull-request in the taxonomy repository that includes the files (e.g. qna.yaml
) with your improved data.
Check out our contributing guide to learn how to contribute.