|
@@ -275,7 +275,7 @@ If you're interested in the field of LLM, you may find the above list of milesto
|
|
|
- [Mesh Tensorflow](https://github.com/tensorflow/mesh) - Mesh TensorFlow: Model Parallelism Made Easier.
|
|
|
- [maxtext](https://github.com/AI-Hypercomputer/maxtext) - A simple, performant and scalable Jax LLM!
|
|
|
- [GPT-NeoX](https://github.com/EleutherAI/gpt-neox) - An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
|
|
|
-
|
|
|
+- [Transformer Engine](https://github.com/NVIDIA/TransformerEngine) - A library for accelerating Transformer model training on NVIDIA GPUs.
|
|
|
|
|
|
## LLM Deployment
|
|
|
|