Matthias Reso
|
b96e435cda
Adjust test_samsum_dataset to second model
|
10 months ago |
Matthias Reso
|
fac41298b0
Adapt test_custom_dataset to new model
|
10 months ago |
Matthias Reso
|
960014a3bb
Fix test_custom_dataset by introducing a stable sort algorithm
|
10 months ago |
Matthias Reso
|
b5583b31d5
Adapt test_grammar_dataset to new model
|
10 months ago |
Matthias Reso
|
17a6d16289
Test batching for both llama versions
|
10 months ago |
Kai Wu
|
7b1a9413d2
fixed a typo
|
10 months ago |
Kai Wu
|
41434dc825
formatted and removed duplicated or unused function get_total_flops() and byte2mb()
|
10 months ago |
Kai Wu
|
f2e80bae22
created a FlopMeasure class on top of FlopCounterMode instead of keep of copy of our own tflop_counter.py
|
10 months ago |
Matthias Reso
|
a414ca6a57
Update chat format for llama3
|
10 months ago |
Kai Wu
|
69e46887b4
handling incorrect profiling early stop caused by max_train_steps and add profiler.step() for each train step
|
10 months ago |
Matthias Reso
|
113ea18bf1
Replace LlamaTokenizer with AutoTokenizer
|
10 months ago |
Beto
|
5979dbe996
Merging local with remote
|
10 months ago |
Kai Wu
|
34e0bf4c6e
second draft of this feature, seems to be working now
|
10 months ago |
Beto
|
d4cbfa1cc1
Merging upstream llama-recipes to current repo
|
10 months ago |
Kai Wu
|
a35519ee90
fixed typo and handling unexpected exit
|
10 months ago |
Kai Wu
|
2a5de9b448
first draft of flop counter feature
|
10 months ago |
Hamid Shojanazeri
|
aaa9e2c863
Adding a feature that will stop the training/eval process after reaching some max_steps (#428)
|
10 months ago |
Kai Wu
|
e6f69f84ad
add max_steps_reached to reduce redundancy
|
10 months ago |
rahul-sarvam
|
0efb8bd31e
Update README.md
|
10 months ago |
rahul-sarvam
|
687c2dc5d8
Update README.md
|
10 months ago |
Rahul A R
|
2fa8e69b62
add new argument: tokenizer_name
|
10 months ago |
Rahul A R
|
f8183b96fe
use new tokenizer_name argument and resize embeddings if required
|
10 months ago |
Rahul A R
|
1e4e3e00fc
adding new multilingual recipe
|
10 months ago |
Kai Wu
|
362cda0fa6
fixing test_gradient_accumulation and test_save_to_json
|
11 months ago |
Kai Wu
|
fa0a389f74
add max_step feature for training and eval
|
11 months ago |
Suraj Subramanian
|
201daff2d1
Add note on CUDA version + remove 'test' from pytorch whl url
|
11 months ago |
Hamid Shojanazeri
|
37c8f72211
Update location and name of llm.py example notebook (#417)
|
11 months ago |
Thomas Robinson
|
79266217ef
Update location and name of llm.py example notebook
|
11 months ago |
Hamid Shojanazeri
|
f7aa02af9f
only save training params on rank 0 (#415)
|
11 months ago |
jpgard
|
6954b16b3b
only save training params on rank 0
|
11 months ago |