merge_mp_bert.sh 794 B

12345678910111213141516171819
  1. #!/bin/bash
  2. TENSOR_MODEL_PARALLEL_SIZE=2
  3. VOCAB_FILE=bert-vocab.txt
  4. CHECKPOINT_PATH=checkpoints/bert_345m
  5. WORLD_SIZE=$TENSOR_MODEL_PARALLEL_SIZE python tools/merge_mp_partitions.py \
  6. --model-type BERT \
  7. --tensor-model-parallel-size $TENSOR_MODEL_PARALLEL_SIZE \
  8. --tokenizer-type BertWordPieceLowerCase \
  9. --vocab-file $VOCAB_FILE \
  10. --num-layers 24 \
  11. --hidden-size 1024 \
  12. --num-attention-heads 16 \
  13. --seq-length 512 \
  14. --max-position-embeddings 512 \
  15. --load $CHECKPOINT_PATH