浏览代码

Merge branch 'main' of github.com:meta-llama/llama-recipes into inference_changes

Beto 8 月之前
父节点
当前提交
4f5fb7c8e9

recipes/use_cases/MediaGen.ipynb → recipes/3p_integrations/octoai/MediaGen.ipynb


+ 5 - 5
recipes/quickstart/README.md

@@ -2,11 +2,11 @@
 
 If you are new to developing with Meta Llama models, this is where you should start. This folder contains introductory-level notebooks across different techniques relating to Meta Llama.
 
-* The [](./Running_Llama3_Anywhere/) notebooks demonstrate how to run Llama inference across Linux, Mac and Windows platforms using the appropriate tooling.
-* The [](./Prompt_Engineering_with_Llama_3.ipynb) notebook showcases the various ways to elicit appropriate outputs from Llama. Take this notebook for a spin to get a feel for how Llama responds to different inputs and generation parameters.
-* The [](./inference/) folder contains scripts to deploy Llama for inference on server and mobile. See also [](../3p_integrations/vllm/) and [](../3p_integrations/tgi/) for hosting Llama on open-source model servers.
-* The [](./RAG/) folder contains a simple Retrieval-Augmented Generation application using Llama 3.
-* The [](./finetuning/) folder contains resources to help you finetune Llama 3 on your custom datasets, for both single- and multi-GPU setups. The scripts use the native llama-recipes finetuning code found in [](../../src/llama_recipes/finetuning.py) which supports these features:
+* The [Running_Llama3_Anywhere](./Running_Llama3_Anywhere/) notebooks demonstrate how to run Llama inference across Linux, Mac and Windows platforms using the appropriate tooling.
+* The [Prompt_Engineering_with_Llama_3](./Prompt_Engineering_with_Llama_3.ipynb) notebook showcases the various ways to elicit appropriate outputs from Llama. Take this notebook for a spin to get a feel for how Llama responds to different inputs and generation parameters.
+* The [inference](./inference/) folder contains scripts to deploy Llama for inference on server and mobile. See also [3p_integrations/vllm](../3p_integrations/vllm/) and [3p_integrations/tgi](../3p_integrations/tgi/) for hosting Llama on open-source model servers.
+* The [RAG](./RAG/) folder contains a simple Retrieval-Augmented Generation application using Llama 3.
+* The [finetuning](./finetuning/) folder contains resources to help you finetune Llama 3 on your custom datasets, for both single- and multi-GPU setups. The scripts use the native llama-recipes finetuning code found in [finetuning.py](../../src/llama_recipes/finetuning.py) which supports these features:
 
 | Feature                                        |   |
 | ---------------------------------------------- | - |

+ 3 - 3
recipes/quickstart/finetuning/LLM_finetuning_overview.md

@@ -33,9 +33,9 @@ Full parameter fine-tuning has its own advantages, in this method there are mult
 You can also keep most of the layers frozen and only fine-tune a few layers. There are many different techniques to choose from to freeze/unfreeze layers based on different criteria.
 
 <div style="display: flex;">
-    <img src="../../docs/img/feature_based_fn.png" alt="Image 1" width="250" />
-    <img src="../../docs/img/feature_based_fn_2.png" alt="Image 2" width="250" />
-    <img src="../../docs/img/full_param_fn.png" alt="Image 3" width="250" />
+    <img src="../../../docs/img/feature_based_fn.png" alt="Image 1" width="250" />
+    <img src="../../../docs/img/feature_based_fn_2.png" alt="Image 2" width="250" />
+    <img src="../../../docs/img/full_param_fn.png" alt="Image 3" width="250" />
 </div>
 
 

+ 1 - 1
recipes/quickstart/finetuning/README.md

@@ -105,7 +105,7 @@ python -m llama_recipes.finetuning --use_peft --peft_method lora --quantization
 ```
 You'll be able to access a dedicated project or run link on [wandb.ai](https://wandb.ai) and see your dashboard like the one below.
 <div style="display: flex;">
-    <img src="../../../docs/images/wandb_screenshot.png" alt="wandb screenshot" width="500" />
+    <img src="../../../docs/img/wandb_screenshot.png" alt="wandb screenshot" width="500" />
 </div>
 
 ## FLOPS Counting and Pytorch Profiling

+ 0 - 3
recipes/use_cases/README.md

@@ -18,6 +18,3 @@ A complete example of how to build a Llama 3 chatbot hosted on your browser that
 
 ## [Sales Bot](./customerservice_chatbots/ai_agent_chatbot/SalesBot.ipynb): Sales Bot with Llama3 - A Summarization and RAG Use Case
 An summarization + RAG use case built around the Amazon product review Kaggle dataset to build a helpful Music Store Sales Bot. The summarization and RAG are built on top of Llama models hosted on OctoAI, and the vector database is hosted on Weaviate Cloud Services.
-
-## [Media Generation](./MediaGen.ipynb): Building a Video Generation Pipeline with Llama3
-This step-by-step tutorial shows how to use leverage Llama 3 to drive the generation of animated videos using SDXL and SVD. More specifically it relies on JSON formatting to produce a scene-by-scene story board of a recipe video. The user provides the name of a dish, then Llama 3 describes a step by step guide to reproduce the said dish. This step by step guide is brought to life with models like SDXL and SVD.