|
1 день тому | |
---|---|---|
.. | ||
data | 3 тижнів тому | |
eval | 1 день тому | |
fine-tuning | 1 день тому | |
quickstart | 2 тижнів тому | |
README.md | 5 днів тому |
This recipe is step by step guide to improve Llama performance on Text2SQL measured with the popular BIRD benchmark. We generate synthetic Chain of Thought(CoT) dataset and fine-tune Llama models on it.
Results: [graph_placeholder]
We followed following steps:
Pre-processing the BIRD TRAIN datset by converting SQL statements into conversation format
We use the conversations from step 1, add CoT to these existing conversations using Llama-3.3-70B
Fine-tuning Llama-3.1-8B on the dataset from step 2
We provide scripts to simplify running the BIRD benchmark on the fine-tuned models and compare it with out of the model.
We also experimented with supervised fine-tuning (SFT) without CoT which resulted in slightly lower accuracy.