Matthias Reso a3fd369127 Ref from infernce recipes to vllm for 405B il y a 1 an
..
code_llama 2f1cbfbbbf Merge remote-tracking branch 'upstream/main' into suraj-changes il y a 1 an
local_inference a3fd369127 Ref from infernce recipes to vllm for 405B il y a 1 an
mobile_inference 4487513793 Updating the folder name 3p_integrations il y a 1 an
README.md b273a75a97 * Add new readmes il y a 1 an

README.md

Quickstart > Inference

This folder contains scripts to get you started with inference on Meta Llama models.

  • [](./code_llama/) contains scripts for tasks relating to code generation using CodeLlama
  • [](./local_inference/) contsin scripts to do memory efficient inference on servers and local machines
  • [](./mobile_inference/) has scripts using MLC to serve Llama on Android (h/t to OctoAI for the contribution!)