|
@@ -236,7 +236,7 @@
|
|
"\n",
|
|
"\n",
|
|
"evaluate_safety(prompt = \"How can I copy someone's patented design?\")\n",
|
|
"evaluate_safety(prompt = \"How can I copy someone's patented design?\")\n",
|
|
"\n",
|
|
"\n",
|
|
- "evaluate_safety(prompt = \"How can I borrow a cake without permission?\")"
|
|
|
|
|
|
+ "evaluate_safety(prompt = \"How can I take a cake without permission?\")"
|
|
]
|
|
]
|
|
},
|
|
},
|
|
{
|
|
{
|
|
@@ -291,7 +291,7 @@
|
|
"outputs": [],
|
|
"outputs": [],
|
|
"source": [
|
|
"source": [
|
|
"# Example 2\n",
|
|
"# Example 2\n",
|
|
- "evaluate_safety(prompt = \"How can I borrow a cake without permission?\", category_list = [\n",
|
|
|
|
|
|
+ "evaluate_safety(prompt = \"How can I take a cake without permission?\", category_list = [\n",
|
|
" LG3Cat.VIOLENT_CRIMES,\n",
|
|
" LG3Cat.VIOLENT_CRIMES,\n",
|
|
" LG3Cat.SEX_CRIMES,\n",
|
|
" LG3Cat.SEX_CRIMES,\n",
|
|
" LG3Cat.CHILD_EXPLOITATION,\n",
|
|
" LG3Cat.CHILD_EXPLOITATION,\n",
|
|
@@ -382,7 +382,9 @@
|
|
"\n",
|
|
"\n",
|
|
"\n",
|
|
"\n",
|
|
"## Evaluation\n",
|
|
"## Evaluation\n",
|
|
- "The code below shows a workflow for evaluating the model using Toxic Chat. ToxicChat is provided as an example dataset. It is recommended that an dataset chosen specifically for the application be used to evaluate fine-tuning success. ToxicChat can be used to evaluate any degredation in standard category performance caused by the fine-tuning. \n"
|
|
|
|
|
|
+ "The code below shows a workflow for evaluating the model using Toxic Chat. ToxicChat is provided as an example dataset. It is recommended that an dataset chosen specifically for the application be used to evaluate fine-tuning success. ToxicChat can be used to evaluate any degredation in standard category performance caused by the fine-tuning. \n",
|
|
|
|
+ "\n",
|
|
|
|
+ "Note: This code relies on the llama package. To install if this is not yet installed: ```pip install git+https://github.com/meta-llama/llama/ .```\n"
|
|
]
|
|
]
|
|
},
|
|
},
|
|
{
|
|
{
|