Ver código fonte

added colab link; some typos?

Jeff Tang 10 meses atrás
pai
commit
e74ba7679f

+ 7 - 4
recipes/finetuning/quickstart_peft_finetuning.ipynb

@@ -6,7 +6,9 @@
    "metadata": {},
    "source": [
     "Copyright (c) Meta Platforms, Inc. and affiliates.\n",
-    "This software may be used and distributed according to the terms of the Llama 2 Community License Agreement."
+    "This software may be used and distributed according to the terms of the Llama 2 Community License Agreement.\n",
+    "\n",
+    "<a href=\"https://colab.research.google.com/github/meta-llama/llama-recipes/blob/main/recipes/finetuning/quickstart_peft_finetuning.ipynb\" target=\"_parent\"><img src=\"https://colab.research.google.com/assets/colab-badge.svg\" alt=\"Open In Colab\"/></a>"
    ]
   },
   {
@@ -18,7 +20,7 @@
     "\n",
     "This notebook shows how to train a Meta Llama 3 model on a single GPU (e.g. A10 with 24GB) using int8 quantization and LoRA finetuning.\n",
     "\n",
-    "**_Note:_** To run this notebook on a machine with less than 24GB VRAM (e.g. T4 with 15GB) the context length of the training dataset needs to be adapted.\n",
+    "**_Note:_** To run this notebook on a machine with less than 24GB VRAM (e.g. T4 with 16GB) the context length of the training dataset needs to be adapted.\n",
     "We do this based on the available VRAM during execution.\n",
     "If you run into OOM issues try to further lower the value of train_config.context_length."
    ]
@@ -38,6 +40,7 @@
    "metadata": {},
    "outputs": [],
    "source": [
+    "# uncomment if running from Colab T4\n",
     "# ! pip install llama-recipes ipywidgets\n",
     "\n",
     "# import huggingface_hub\n",
@@ -95,7 +98,7 @@
     "train_config.lr = 3e-4\n",
     "train_config.use_fast_kernels = True\n",
     "train_config.use_fp16 = True\n",
-    "train_config.context_length = 1024 if torch.cuda.get_device_properties(0).total_memory < 16e9 else 2048 # T4 15GB or A10 24GB\n",
+    "train_config.context_length = 1024 if torch.cuda.get_device_properties(0).total_memory < 16e9 else 2048 # T4 16GB or A10 24GB\n",
     "train_config.batching_strategy = \"packing\"\n",
     "train_config.output_dir = \"meta-llama-samsum\"\n",
     "\n",
@@ -464,7 +467,7 @@
    "name": "python",
    "nbconvert_exporter": "python",
    "pygments_lexer": "ipython3",
-   "version": "3.11.9"
+   "version": "3.10.14"
   },
   "vscode": {
    "interpreter": {