README.md 1.1 KB

Convert Hugging Face llama weights to official llama consolidated format

This is the reverse conversion for convert_llama_weights_to_hf.py script from the transformer package.

Step 0: Convert to consolidated format

  • Create an output directory for the converted weights, such as test70B.
  • Copy file params.json from the official llama download into that directory.
  • Run the conversion script. model-path can be a Hugging Face hub model or a local hf model directory. python -m llama_recipes.tools.convert_hf_weights_to_llama --model-path meta-llama/Meta-Llama-3.1-70B-Instruct --output-dir test70B --model-size 70B

Step 1: Run inference

Checkout the official llama 3 inference repo. Test using chat or text completion.

torchrun --nproc_per_node 8 example_chat_completion.py --ckpt_dir ./test70B --tokenizer_path ${llama_3_dir}/tokenizer.model

For validation, please compare the converted weights with official llama 2 weights

python compare_llama_weights.py test70B ${Llama-3-70B-Instruct_dir}