Skip to content

v0.6.0: LLaMa (Alpaca), Benchmark Util, T5 ILQL, Tests

Compare
Choose a tag to compare
@jon-tow jon-tow released this 31 Mar 21:41
b7db6f9

The v0.6.0 release includes several new features, bug fixes, and overall improvements to the codebase. Here are the key changes:

📏 Benchmarking and Improved Unit Tests

This release introduces a new benchmark util to more easily track regressions in our training pipeline along with improved unit tests with the help of the hypothesis package:

🦙 LLaMa and Alpaca PPO/SFT Support

PPO support and examples for LLaMa are now available and we’ve baked in an example for instruction fine-tuning models with the Alpaca dataset using our SFT trainer:

5️⃣ T5 ILQL Support

T5 models can now be fine-tuned with ILQL:

  • Support ILQL for T5 model, Fix PPO T5 for refactored code by @PhungVanDuy in #290

Fixes

What's Changed

  • Move to Python config classes instead of ymls by @cat-state in #306
  • Add intermediate checkpointing to accelerate trainers by @jon-tow in #349
  • Enable infinite dataloader for prompt_dataloader in PPO Trainer by @alexandremuzio in #358
  • [feat] Add optional dependency list by @reciprocated in #381
  • Add some synchronization to the db download in the simulacra example by @dakinggg in #406

New Contributors

Full Changelog: v0.5.0...v0.6.0