瀏覽代碼

Merge pull request #189 from vg11072001/patch-1

Transformer Engine lib added on llm training
Hannibal046 4 月之前
父節點
當前提交
ec22eaa298
共有 1 個文件被更改,包括 1 次插入1 次删除
  1. 1 1
      README.md

+ 1 - 1
README.md

@@ -275,7 +275,7 @@ If you're interested in the field of LLM, you may find the above list of milesto
 - [Mesh Tensorflow](https://github.com/tensorflow/mesh) - Mesh TensorFlow: Model Parallelism Made Easier.
 - [maxtext](https://github.com/AI-Hypercomputer/maxtext) - A simple, performant and scalable Jax LLM!
 - [GPT-NeoX](https://github.com/EleutherAI/gpt-neox) - An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
-
+- [Transformer Engine](https://github.com/NVIDIA/TransformerEngine) - A library for accelerating Transformer model training on NVIDIA GPUs.
 
 ## LLM Deployment