Reimplementation of the task generation part from the Alpaca paper
🦙🗽 Small research project - how much it would cost to create Alpaca-like dataset, with 50k+ demonstrations, using slightly different approach. All data byproducts are CC0/MIT-licensed.
🔥 The project also contains 100k+ MIT-licensed demonstrations from Anthropics HH-RLHF repo - converted into "Alpaca compatible format".
👉 Follow me on Twitter for news and updates.
🚫 Remember that releasing a model based on data you generated via model API might violate the Terms of Service of the model API provider.
BTW: This repo shows how easy it is to fine-tune (PEFT=LORA) Flan-T5-* model with Alpaca-like dataset.
git clone https://github.com/mobarski/alpaca-libre && cd alpaca-libre
pip install -r requirements.txt
export OPENAI_KEY=...
python3 generate.py
data/seed_tasks.jsonl
- is from the Self-Instruct paperdata/alpaca_libre_prompt_v1.txt
- is from the Alpaca paper (with slight modfification)Files in the data/output
directory are in the same format as original Alpaca dataset.
Files in the data/output/work
directory are in the .jsonl format and:
contain one task (JSON object) per line,
contain also tasks that failed quality checks (status!='ok')
each task object has the following items:
status - anything other than 'ok' is bad
instruction - instruction part of the prompt
input - input part of the prompt
output - expected output
other - dictionary for other information (similarity, etc)
GitHub repos:
Papers:
<noinput>
handling<nooutput>
handling<noinput>