|
|
1 yıl önce | |
|---|---|---|
| .. | ||
| hf_text_generation_inference | 6d449a859b New folder structure (#1) | 2 yıl önce |
| vllm | 6d449a859b New folder structure (#1) | 2 yıl önce |
| README.md | 6d449a859b New folder structure (#1) | 2 yıl önce |
| llama-on-prem.md | 5cedef2155 Update llama-on-prem.md | 1 yıl önce |
This tutorial shows how to use Llama 2 with vLLM and Hugging Face TGI to build Llama 2 on-prem apps.
* To run a quantized Llama2 model on iOS and Android, you can use the open source MLC LLM or llama.cpp. You can even make a Linux OS that boots to Llama2 (repo).