|
@@ -145,7 +145,7 @@
|
|
|
"class Args:\n",
|
|
|
" def __init__(self, \n",
|
|
|
" max_examples=100, \n",
|
|
|
- " sql_model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\", \n",
|
|
|
+ " sql_model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\", \n",
|
|
|
" gold_file_name=\"gold-test-set.jsonl\",\n",
|
|
|
" training_file_name=\"generated_queries.jsonl\",\n",
|
|
|
" num_to_generate=10):\n",
|
|
@@ -197,7 +197,7 @@
|
|
|
}
|
|
|
],
|
|
|
"source": [
|
|
|
- "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\")\n",
|
|
|
+ "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\")\n",
|
|
|
"\n",
|
|
|
"question = \"\"\"Who is the highest paid NBA player?\"\"\"\n",
|
|
|
"system = f\"\"\"You are an NBA analyst with 15 years of experience writing complex SQL queries. Consider the nba_roster table with the following schema:\n",
|
|
@@ -418,7 +418,7 @@
|
|
|
"class ScoreStage(GenerationNode):\n",
|
|
|
" def __init__(self):\n",
|
|
|
" super().__init__(\n",
|
|
|
- " model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\",\n",
|
|
|
+ " model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\",\n",
|
|
|
" max_new_tokens=150,\n",
|
|
|
" )\n",
|
|
|
"\n",
|
|
@@ -712,7 +712,7 @@
|
|
|
"class ModelStage(GenerationNode):\n",
|
|
|
" def __init__(self):\n",
|
|
|
" super().__init__(\n",
|
|
|
- " model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\",\n",
|
|
|
+ " model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\",\n",
|
|
|
" max_new_tokens=300,\n",
|
|
|
" )\n",
|
|
|
"\n",
|
|
@@ -808,7 +808,7 @@
|
|
|
"class QuestionStage(GenerationNode):\n",
|
|
|
" def __init__(self):\n",
|
|
|
" super().__init__(\n",
|
|
|
- " model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\",\n",
|
|
|
+ " model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\",\n",
|
|
|
" max_new_tokens=150,\n",
|
|
|
" )\n",
|
|
|
"\n",
|
|
@@ -1055,7 +1055,7 @@
|
|
|
],
|
|
|
"source": [
|
|
|
"args = Args()\n",
|
|
|
- "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\")\n",
|
|
|
+ "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\")\n",
|
|
|
"\n",
|
|
|
"dataset = get_dataset(args, make_question)\n",
|
|
|
"finetune_args = get_default_finetune_args()\n",
|
|
@@ -1601,7 +1601,7 @@
|
|
|
],
|
|
|
"source": [
|
|
|
"args = Args(training_file_name=\"archive/generated_queries_large_filtered_cleaned.jsonl\")\n",
|
|
|
- "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\")\n",
|
|
|
+ "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\")\n",
|
|
|
"\n",
|
|
|
"dataset = get_dataset(args, make_question)\n",
|
|
|
"finetune_args = get_default_finetune_args()\n",
|
|
@@ -1798,7 +1798,7 @@
|
|
|
],
|
|
|
"source": [
|
|
|
"args = Args(training_file_name=\"generated_queries_v2.jsonl\")\n",
|
|
|
- "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\")\n",
|
|
|
+ "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\")\n",
|
|
|
"\n",
|
|
|
"dataset = get_dataset(args, make_question)\n",
|
|
|
"finetune_args = get_default_finetune_args()\n",
|
|
@@ -1966,7 +1966,7 @@
|
|
|
],
|
|
|
"source": [
|
|
|
"args = Args(training_file_name=\"archive/generated_queries_v2_large_filtered_cleaned.jsonl\")\n",
|
|
|
- "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3-8B-Instruct\")\n",
|
|
|
+ "llm = lamini.Lamini(model_name=\"meta-llama/Meta-Llama-3.1-8B-Instruct\")\n",
|
|
|
"\n",
|
|
|
"dataset = get_dataset(args, make_question)\n",
|
|
|
"finetune_args = get_default_finetune_args()\n",
|