|
1 周之前 | |
---|---|---|
.. | ||
data | 3 周之前 | |
eval | 1 周之前 | |
fine-tuning | 1 周之前 | |
quickstart | 2 周之前 | |
README.md | 1 周之前 |
This folder contains scripts to:
Evaluate Llama (original and fine-tuned) models on the Text2SQL task using the popular BIRD dataset.
Generate two supervised fine-tuning (SFT) datasets (with and without CoT) and fine-tuning Llama 3.1 8B with the datasets, using different SFT options: with or without CoT, using quantization or not, full fine-tuning (FFT) or parameter-efficient fine-tuning (PEFT). The non-quantized PEFT CoT SFT has the most performance gains: from 39.47% of the original Llama 3.1 8B model to 43.35%. (Note: the results are based on 3 epochs of SFT.)
Our end goal is to maximize the accuracy of Llama models on the Text2SQL task. To do so we need to first evaluate the current state of the art Llama models on the task, then apply fine-tuning, agent and other approaches to evaluate and improve Llama's performance.