You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This issue captures some of the key steps required to reproduce the Constitutional AI paper steps to fine tune a RLHF model with feedback generated by a RLAIF model.
Phase One
Gather a dataset of harmful prompts
Create a base script to compose prompts using a base constitution
Generate a new dataset of prompts + responses using Carper's GPT-J RLHF to review / critique the output
Fine-tune the original model on revised responses using supervised learning
Phase Two
Sample the fine tuned model using the dataset of harmful prompts to create a new dataset with multiple outputs
Overview
This issue captures some of the key steps required to reproduce the Constitutional AI paper steps to fine tune a RLHF model with feedback generated by a RLAIF model.
Phase One
Phase Two
The text was updated successfully, but these errors were encountered: