Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Integration of Self-Play Fine-Tuning (SPIN) Method for Enhancing Large Language Models #588

Open
SeungyounShin opened this issue Jan 11, 2024 · 0 comments
Labels
feature request New feature or request

Comments

@SeungyounShin
Copy link

SeungyounShin commented Jan 11, 2024

🚀 The feature, motivation, and pitch

The recent paper, "Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models," presents a novel method called Self-Play fIne-tuNing (SPIN). This method significantly improves the performance of Large Language Models (LLMs) without additional human preference or AI-feedback data. I'm currently working on language model enhancement and believe integrating SPIN into the TRX1 library could greatly benefit the community. SPIN starts with a supervised fine-tuned model and utilizes self-play, where the LLM refines its capabilities by playing against instances of itself. This approach allows LLMs to generate their own training data from previous iterations, discerning these self-generated responses from human-annotated data, thus progressively improving the model. The integration of SPIN into TRX would enable researchers and developers to easily enhance their LLMs, potentially achieving human-level performance without the need for extensive annotated datasets.

Arxiv

Alternatives

x

Additional context

The SPIN method has been theoretically proven and empirically evaluated on several benchmarks, including the HuggingFace Open LLM Leaderboard, MT-Bench, and datasets from Big-Bench. The results show that SPIN can significantly improve LLM performance across a variety of tasks, even outperforming models trained through direct preference optimization supplemented with extra GPT-4 preference data. Integrating this method into the TRX could open up new possibilities for enhancing LLMs efficiently.

@SeungyounShin SeungyounShin added the feature request New feature or request label Jan 11, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature request New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant