Parcourir la source

Merge pull request #174 from bitxsw93/feature/mnn_llm

[LLM Inference] MNN-LLM: llm on device inference framework
Hannibal046 il y a 5 mois
Parent
commit
bbcd0295b3
1 fichiers modifiés avec 1 ajouts et 0 suppressions
  1. 1 0
      README.md

+ 1 - 0
README.md

@@ -325,6 +325,7 @@ If you're interested in the field of LLM, you may find the above list of milesto
 - [talkd.ai dialog](https://github.com/talkdai/dialog) - Simple API for deploying any RAG or LLM that you want adding plugins.
 - [Wllama](https://github.com/ngxson/wllama) - WebAssembly binding for llama.cpp - Enabling in-browser LLM inference
 - [GPUStack](https://github.com/gpustack/gpustack) - An open-source GPU cluster manager for running LLMs
+- [MNN-LLM](https://github.com/alibaba/MNN) -- A Device-Inference framework, including LLM Inference on device(Mobile Phone/PC/IOT)
 
 ## LLM Applications
 - [AdalFlow](https://github.com/SylphAI-Inc/AdalFlow) - AdalFlow: The PyTorch library for LLM applications.