|  rahul-sarvam | eb7ef4225f
							
							Update recipes/multilingual/README.md | 1 年之前 | 
				
					
						|  rahul-sarvam | f1f335a591
							
							Update recipes/multilingual/README.md | 1 年之前 | 
				
					
						|  rahul-sarvam | 47556ce0a6
							
							Update recipes/multilingual/README.md | 1 年之前 | 
				
					
						|  Matthias Reso | 739483f262
							
							Adjust test_grammar_datasets to stable sort | 1 年之前 | 
				
					
						|  Matthias Reso | b96e435cda
							
							Adjust test_samsum_dataset to second model | 1 年之前 | 
				
					
						|  Matthias Reso | fac41298b0
							
							Adapt test_custom_dataset to new model | 1 年之前 | 
				
					
						|  Matthias Reso | 960014a3bb
							
							Fix test_custom_dataset by introducing a stable sort algorithm | 1 年之前 | 
				
					
						|  Matthias Reso | b5583b31d5
							
							Adapt test_grammar_dataset to new model | 1 年之前 | 
				
					
						|  Matthias Reso | 17a6d16289
							
							Test batching for both llama versions | 1 年之前 | 
				
					
						|  Kai Wu | 7b1a9413d2
							
							fixed a typo | 1 年之前 | 
				
					
						|  Kai Wu | 41434dc825
							
							formatted and removed duplicated or unused function get_total_flops() and byte2mb() | 1 年之前 | 
				
					
						|  Kai Wu | f2e80bae22
							
							created a FlopMeasure class on top of FlopCounterMode instead of keep of copy of our own tflop_counter.py | 1 年之前 | 
				
					
						|  Matthias Reso | a414ca6a57
							
							Update chat format for llama3 | 1 年之前 | 
				
					
						|  Kai Wu | 69e46887b4
							
							handling incorrect profiling early stop caused by max_train_steps and add profiler.step() for each train step | 1 年之前 | 
				
					
						|  Matthias Reso | 113ea18bf1
							
							Replace LlamaTokenizer with AutoTokenizer | 1 年之前 | 
				
					
						|  Beto | 5979dbe996
							
							Merging local with remote | 1 年之前 | 
				
					
						|  Kai Wu | 34e0bf4c6e
							
							second draft of this feature, seems to be working now | 1 年之前 | 
				
					
						|  Beto | d4cbfa1cc1
							
							Merging upstream llama-recipes to current repo | 1 年之前 | 
				
					
						|  Kai Wu | a35519ee90
							
							fixed typo and handling unexpected exit | 1 年之前 | 
				
					
						|  Kai Wu | 2a5de9b448
							
							first draft of flop counter feature | 1 年之前 | 
				
					
						|  Hamid Shojanazeri | aaa9e2c863
							
							Adding a feature that will stop the training/eval process after reaching some max_steps (#428) | 1 年之前 | 
				
					
						|  Kai Wu | e6f69f84ad
							
							add max_steps_reached to reduce redundancy | 1 年之前 | 
				
					
						|  rahul-sarvam | 0efb8bd31e
							
							Update README.md | 1 年之前 | 
				
					
						|  rahul-sarvam | 687c2dc5d8
							
							Update README.md | 1 年之前 | 
				
					
						|  Rahul A R | 2fa8e69b62
							
							add new argument: tokenizer_name | 1 年之前 | 
				
					
						|  Rahul A R | f8183b96fe
							
							use new tokenizer_name argument and resize embeddings if required | 1 年之前 | 
				
					
						|  Rahul A R | 1e4e3e00fc
							
							adding new multilingual recipe | 1 年之前 | 
				
					
						|  Kai Wu | 362cda0fa6
							
							fixing test_gradient_accumulation and test_save_to_json | 1 年之前 | 
				
					
						|  Kai Wu | fa0a389f74
							
							add max_step feature for training and eval | 1 年之前 | 
				
					
						|  Suraj Subramanian | 201daff2d1
							
							Add note on CUDA version + remove 'test' from pytorch whl url | 1 年之前 |