浏览代码

Merge pull request #12 from meta-llama/upstream_merge

Upstream merge into alpha main
Suraj Subramanian 11 月之前
父节点
当前提交
a1d51de4b0

recipes/use_cases/MediaGen.ipynb → recipes/3p_integrations/octoai/MediaGen.ipynb


+ 5 - 5
recipes/quickstart/README.md

@@ -2,11 +2,11 @@
 
 
 If you are new to developing with Meta Llama models, this is where you should start. This folder contains introductory-level notebooks across different techniques relating to Meta Llama.
 If you are new to developing with Meta Llama models, this is where you should start. This folder contains introductory-level notebooks across different techniques relating to Meta Llama.
 
 
-* The [](./Running_Llama3_Anywhere/) notebooks demonstrate how to run Llama inference across Linux, Mac and Windows platforms using the appropriate tooling.
-* The [](./Prompt_Engineering_with_Llama_3.ipynb) notebook showcases the various ways to elicit appropriate outputs from Llama. Take this notebook for a spin to get a feel for how Llama responds to different inputs and generation parameters.
-* The [](./inference/) folder contains scripts to deploy Llama for inference on server and mobile. See also [](../3p_integrations/vllm/) and [](../3p_integrations/tgi/) for hosting Llama on open-source model servers.
-* The [](./RAG/) folder contains a simple Retrieval-Augmented Generation application using Llama 3.
-* The [](./finetuning/) folder contains resources to help you finetune Llama 3 on your custom datasets, for both single- and multi-GPU setups. The scripts use the native llama-recipes finetuning code found in [](../../src/llama_recipes/finetuning.py) which supports these features:
+* The [Running_Llama3_Anywhere](./Running_Llama3_Anywhere/) notebooks demonstrate how to run Llama inference across Linux, Mac and Windows platforms using the appropriate tooling.
+* The [Prompt_Engineering_with_Llama_3](./Prompt_Engineering_with_Llama_3.ipynb) notebook showcases the various ways to elicit appropriate outputs from Llama. Take this notebook for a spin to get a feel for how Llama responds to different inputs and generation parameters.
+* The [inference](./inference/) folder contains scripts to deploy Llama for inference on server and mobile. See also [3p_integrations/vllm](../3p_integrations/vllm/) and [3p_integrations/tgi](../3p_integrations/tgi/) for hosting Llama on open-source model servers.
+* The [RAG](./RAG/) folder contains a simple Retrieval-Augmented Generation application using Llama 3.
+* The [finetuning](./finetuning/) folder contains resources to help you finetune Llama 3 on your custom datasets, for both single- and multi-GPU setups. The scripts use the native llama-recipes finetuning code found in [finetuning.py](../../src/llama_recipes/finetuning.py) which supports these features:
 
 
 | Feature                                        |   |
 | Feature                                        |   |
 | ---------------------------------------------- | - |
 | ---------------------------------------------- | - |

+ 3 - 3
recipes/quickstart/finetuning/LLM_finetuning_overview.md

@@ -33,9 +33,9 @@ Full parameter fine-tuning has its own advantages, in this method there are mult
 You can also keep most of the layers frozen and only fine-tune a few layers. There are many different techniques to choose from to freeze/unfreeze layers based on different criteria.
 You can also keep most of the layers frozen and only fine-tune a few layers. There are many different techniques to choose from to freeze/unfreeze layers based on different criteria.
 
 
 <div style="display: flex;">
 <div style="display: flex;">
-    <img src="../../docs/img/feature_based_fn.png" alt="Image 1" width="250" />
-    <img src="../../docs/img/feature_based_fn_2.png" alt="Image 2" width="250" />
-    <img src="../../docs/img/full_param_fn.png" alt="Image 3" width="250" />
+    <img src="../../../docs/img/feature_based_fn.png" alt="Image 1" width="250" />
+    <img src="../../../docs/img/feature_based_fn_2.png" alt="Image 2" width="250" />
+    <img src="../../../docs/img/full_param_fn.png" alt="Image 3" width="250" />
 </div>
 </div>
 
 
 
 

+ 1 - 1
recipes/quickstart/finetuning/README.md

@@ -105,7 +105,7 @@ python -m llama_recipes.finetuning --use_peft --peft_method lora --quantization
 ```
 ```
 You'll be able to access a dedicated project or run link on [wandb.ai](https://wandb.ai) and see your dashboard like the one below.
 You'll be able to access a dedicated project or run link on [wandb.ai](https://wandb.ai) and see your dashboard like the one below.
 <div style="display: flex;">
 <div style="display: flex;">
-    <img src="../../../docs/images/wandb_screenshot.png" alt="wandb screenshot" width="500" />
+    <img src="../../../docs/img/wandb_screenshot.png" alt="wandb screenshot" width="500" />
 </div>
 </div>
 
 
 ## FLOPS Counting and Pytorch Profiling
 ## FLOPS Counting and Pytorch Profiling

+ 0 - 3
recipes/use_cases/README.md

@@ -18,6 +18,3 @@ A complete example of how to build a Llama 3 chatbot hosted on your browser that
 
 
 ## [Sales Bot](./customerservice_chatbots/ai_agent_chatbot/SalesBot.ipynb): Sales Bot with Llama3 - A Summarization and RAG Use Case
 ## [Sales Bot](./customerservice_chatbots/ai_agent_chatbot/SalesBot.ipynb): Sales Bot with Llama3 - A Summarization and RAG Use Case
 An summarization + RAG use case built around the Amazon product review Kaggle dataset to build a helpful Music Store Sales Bot. The summarization and RAG are built on top of Llama models hosted on OctoAI, and the vector database is hosted on Weaviate Cloud Services.
 An summarization + RAG use case built around the Amazon product review Kaggle dataset to build a helpful Music Store Sales Bot. The summarization and RAG are built on top of Llama models hosted on OctoAI, and the vector database is hosted on Weaviate Cloud Services.
-
-## [Media Generation](./MediaGen.ipynb): Building a Video Generation Pipeline with Llama3
-This step-by-step tutorial shows how to use leverage Llama 3 to drive the generation of animated videos using SDXL and SVD. More specifically it relies on JSON formatting to produce a scene-by-scene story board of a recipe video. The user provides the name of a dish, then Llama 3 describes a step by step guide to reproduce the said dish. This step by step guide is brought to life with models like SDXL and SVD.