This website works better with JavaScript
Startseite
Erkunden
Hilfe
Anmelden
radu
/
LLamaRecipes
Mirror von
https://github.com/facebookresearch/llama-recipes.git
Beobachten
1
Favorit hinzufügen
0
Fork
0
Dateien
Issues
0
Wiki
Quellcode durchsuchen
update readme
Kai Wu
vor 7 Monaten
Ursprung
a7b449234a
Commit
576e574e31
17 geänderte Dateien
mit
26 neuen
und
26 gelöschten
Zeilen
Geteilte Ansicht
Diff-Statistik anzeigen
2
2
tools/benchmarks/llm_eval_harness/README.md
24
24
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/README.md
0
0
tools/benchmarks/llm_eval_harness/meta_eval/eval_config.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/bbh/bbh_3shot_cot.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/bbh/utils.py
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/gpqa_cot/gpqa_0shot_cot.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/gpqa_cot/utils.py
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/ifeval/ifeval.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/ifeval/utils.py
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/math_hard/math_hard_0shot_cot.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/math_hard/utils.py
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/meta_instruct.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/meta_pretrain.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/mmlu_pro/mmlu_pro_5shot_cot_instruct.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/mmlu_pro/mmlu_pro_5shot_cot_pretrain.yaml
0
0
tools/benchmarks/llm_eval_harness/meta_eval/meta_template/mmlu_pro/utils.py
0
0
tools/benchmarks/llm_eval_harness/meta_eval/prepare_meta_eval.py
Datei-Diff unterdrückt, da er zu groß ist
+ 2
- 2
tools/benchmarks/llm_eval_harness/README.md
Datei-Diff unterdrückt, da er zu groß ist
+ 24
- 24
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/README.md
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/eval_config.yaml → tools/benchmarks/llm_eval_harness/meta_eval/eval_config.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/bbh/bbh_3shot_cot.yaml → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/bbh/bbh_3shot_cot.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/bbh/utils.py → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/bbh/utils.py
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/gpqa_cot/gpqa_0shot_cot.yaml → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/gpqa_cot/gpqa_0shot_cot.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/gpqa_cot/utils.py → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/gpqa_cot/utils.py
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/ifeval/ifeval.yaml → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/ifeval/ifeval.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/ifeval/utils.py → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/ifeval/utils.py
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/math_hard/math_hard_0shot_cot.yaml → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/math_hard/math_hard_0shot_cot.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/math_hard/utils.py → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/math_hard/utils.py
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/meta_instruct.yaml → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/meta_instruct.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/meta_pretrain.yaml → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/meta_pretrain.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/mmlu_pro/mmlu_pro_5shot_cot_instruct.yaml → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/mmlu_pro/mmlu_pro_5shot_cot_instruct.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/mmlu_pro/mmlu_pro_5shot_cot_pretrain.yaml → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/mmlu_pro/mmlu_pro_5shot_cot_pretrain.yaml
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/meta_template/mmlu_pro/utils.py → tools/benchmarks/llm_eval_harness/meta_eval/meta_template/mmlu_pro/utils.py
Datei anzeigen
tools/benchmarks/llm_eval_harness/meta_eval_reproduce/prepare_meta_eval.py → tools/benchmarks/llm_eval_harness/meta_eval/prepare_meta_eval.py
Datei anzeigen