Browse Source

Updated tools folder under llama-recipes

Pia Papanna 9 months ago
parent
commit
6ff1184120
32 changed files with 1 additions and 93451 deletions
  1. 1 1
      .github/scripts/check_copyright_header.py
  2. 0 1
      recipes/README.md
  3. 0 23
      recipes/benchmarks/inference_throughput/tokenizer/special_tokens_map.json
  4. 0 93391
      recipes/benchmarks/inference_throughput/tokenizer/tokenizer.json
  5. BIN
      recipes/benchmarks/inference_throughput/tokenizer/tokenizer.model
  6. 0 35
      recipes/benchmarks/inference_throughput/tokenizer/tokenizer_config.json
  7. 0 0
      tools/benchmarks/inference/README.md
  8. 0 0
      tools/benchmarks/inference/cloud/README.md
  9. 0 0
      tools/benchmarks/inference/cloud/aws/fmbench/README.md
  10. 0 0
      tools/benchmarks/inference/cloud/aws/fmbench/config.yml
  11. 0 0
      tools/benchmarks/inference/cloud/aws/fmbench/img/CFT.png
  12. 0 0
      tools/benchmarks/inference/cloud/aws/fmbench/img/business_summary.png
  13. 0 0
      tools/benchmarks/inference/cloud/aws/fmbench/img/instances.png
  14. 0 0
      tools/benchmarks/inference/cloud/aws/fmbench/img/latency_vs_tokens.png
  15. 0 0
      tools/benchmarks/inference/cloud/azure/chat_azure_api_benchmark.py
  16. 0 0
      tools/benchmarks/inference/cloud/azure/input.jsonl
  17. 0 0
      tools/benchmarks/inference/cloud/azure/parameters.json
  18. 0 0
      tools/benchmarks/inference/cloud/azure/pretrained_azure_api_benchmark.py
  19. 0 0
      tools/benchmarks/inference/on_prem/README.md
  20. 0 0
      tools/benchmarks/inference/on_prem/vllm/chat_vllm_benchmark.py
  21. 0 0
      tools/benchmarks/inference/on_prem/vllm/input.jsonl
  22. 0 0
      tools/benchmarks/inference/on_prem/vllm/parameters.json
  23. 0 0
      tools/benchmarks/inference/on_prem/vllm/pretrained_vllm_benchmark.py
  24. 0 0
      tools/benchmarks/inference/requirements.txt
  25. 0 0
      tools/benchmarks/llm_eval_harness/README.md
  26. 0 0
      tools/benchmarks/llm_eval_harness/eval.py
  27. 0 0
      tools/benchmarks/llm_eval_harness/open_llm_eval_prep.sh
  28. 0 0
      tools/benchmarks/llm_eval_harness/open_llm_leaderboard/arc_challeneg_25shots.yaml
  29. 0 0
      tools/benchmarks/llm_eval_harness/open_llm_leaderboard/hellaswag_10shots.yaml
  30. 0 0
      tools/benchmarks/llm_eval_harness/open_llm_leaderboard/hellaswag_utils.py
  31. 0 0
      tools/benchmarks/llm_eval_harness/open_llm_leaderboard/mmlu_5shots.yaml
  32. 0 0
      tools/benchmarks/llm_eval_harness/open_llm_leaderboard/winogrande_5shots.yaml

+ 1 - 1
.github/scripts/check_copyright_header.py

@@ -11,7 +11,7 @@ HEADER = """# Copyright (c) Meta Platforms, Inc. and affiliates.
 # This software may be used and distributed according to the terms of the Llama 2 Community License Agreement.\n\n"""
 
 #Files in black list must be relative to main repo folder
-BLACKLIST = ["eval/open_llm_leaderboard/hellaswag_utils.py"]
+BLACKLIST = ["tools/benchmarks/llm_eval_harness/open_llm_leaderboard/hellaswag_utils.py"]
 
 if __name__ == "__main__":
     for ext in ["*.py", "*.sh"]:

+ 0 - 1
recipes/README.md

@@ -11,4 +11,3 @@ This folder contains examples organized by topic:
 [llama_api_providers](./llama_api_providers)|Scripts to run inference on Llama via hosted endpoints
 [benchmarks](./benchmarks)|Scripts to benchmark Llama 3 models inference on various backends
 [code_llama](./code_llama)|Scripts to run inference with the Code Llama models
-[evaluation](./evaluation)|Scripts to evaluate fine-tuned Llama 3 models using `lm-evaluation-harness` from `EleutherAI`

+ 0 - 23
recipes/benchmarks/inference_throughput/tokenizer/special_tokens_map.json

@@ -1,23 +0,0 @@
-{
-  "bos_token": {
-    "content": "<s>",
-    "lstrip": false,
-    "normalized": true,
-    "rstrip": false,
-    "single_word": false
-  },
-  "eos_token": {
-    "content": "</s>",
-    "lstrip": false,
-    "normalized": true,
-    "rstrip": false,
-    "single_word": false
-  },
-  "unk_token": {
-    "content": "<unk>",
-    "lstrip": false,
-    "normalized": true,
-    "rstrip": false,
-    "single_word": false
-  }
-}

File diff suppressed because it is too large
+ 0 - 93391
recipes/benchmarks/inference_throughput/tokenizer/tokenizer.json


BIN
recipes/benchmarks/inference_throughput/tokenizer/tokenizer.model


+ 0 - 35
recipes/benchmarks/inference_throughput/tokenizer/tokenizer_config.json

@@ -1,35 +0,0 @@
-{
-  "add_bos_token": true,
-  "add_eos_token": false,
-  "bos_token": {
-    "__type": "AddedToken",
-    "content": "<s>",
-    "lstrip": false,
-    "normalized": true,
-    "rstrip": false,
-    "single_word": false
-  },
-  "clean_up_tokenization_spaces": false,
-  "eos_token": {
-    "__type": "AddedToken",
-    "content": "</s>",
-    "lstrip": false,
-    "normalized": true,
-    "rstrip": false,
-    "single_word": false
-  },
-  "legacy": true,
-  "use_default_system_prompt": false,
-  "model_max_length": 1000000000000000019884624838656,
-  "pad_token": null,
-  "sp_model_kwargs": {},
-  "tokenizer_class": "LlamaTokenizerFast",
-  "unk_token": {
-    "__type": "AddedToken",
-    "content": "<unk>",
-    "lstrip": false,
-    "normalized": true,
-    "rstrip": false,
-    "single_word": false
-  }
-}

recipes/benchmarks/inference_throughput/README.md → tools/benchmarks/inference/README.md


recipes/benchmarks/inference_throughput/cloud-api/README.md → tools/benchmarks/inference/cloud/README.md


recipes/benchmarks/fmbench/README.md → tools/benchmarks/inference/cloud/aws/fmbench/README.md


recipes/benchmarks/fmbench/config.yml → tools/benchmarks/inference/cloud/aws/fmbench/config.yml


recipes/benchmarks/fmbench/img/CFT.png → tools/benchmarks/inference/cloud/aws/fmbench/img/CFT.png


recipes/benchmarks/fmbench/img/business_summary.png → tools/benchmarks/inference/cloud/aws/fmbench/img/business_summary.png


recipes/benchmarks/fmbench/img/instances.png → tools/benchmarks/inference/cloud/aws/fmbench/img/instances.png


recipes/benchmarks/fmbench/img/latency_vs_tokens.png → tools/benchmarks/inference/cloud/aws/fmbench/img/latency_vs_tokens.png


recipes/benchmarks/inference_throughput/cloud-api/azure/chat_azure_api_benchmark.py → tools/benchmarks/inference/cloud/azure/chat_azure_api_benchmark.py


recipes/benchmarks/inference_throughput/cloud-api/azure/input.jsonl → tools/benchmarks/inference/cloud/azure/input.jsonl


recipes/benchmarks/inference_throughput/cloud-api/azure/parameters.json → tools/benchmarks/inference/cloud/azure/parameters.json


recipes/benchmarks/inference_throughput/cloud-api/azure/pretrained_azure_api_benchmark.py → tools/benchmarks/inference/cloud/azure/pretrained_azure_api_benchmark.py


recipes/benchmarks/inference_throughput/on-prem/README.md → tools/benchmarks/inference/on_prem/README.md


recipes/benchmarks/inference_throughput/on-prem/vllm/chat_vllm_benchmark.py → tools/benchmarks/inference/on_prem/vllm/chat_vllm_benchmark.py


recipes/benchmarks/inference_throughput/on-prem/vllm/input.jsonl → tools/benchmarks/inference/on_prem/vllm/input.jsonl


recipes/benchmarks/inference_throughput/on-prem/vllm/parameters.json → tools/benchmarks/inference/on_prem/vllm/parameters.json


recipes/benchmarks/inference_throughput/on-prem/vllm/pretrained_vllm_benchmark.py → tools/benchmarks/inference/on_prem/vllm/pretrained_vllm_benchmark.py


recipes/benchmarks/inference_throughput/requirements.txt → tools/benchmarks/inference/requirements.txt


recipes/evaluation/README.md → tools/benchmarks/llm_eval_harness/README.md


recipes/evaluation/eval.py → tools/benchmarks/llm_eval_harness/eval.py


recipes/evaluation/open_llm_eval_prep.sh → tools/benchmarks/llm_eval_harness/open_llm_eval_prep.sh


recipes/evaluation/open_llm_leaderboard/arc_challeneg_25shots.yaml → tools/benchmarks/llm_eval_harness/open_llm_leaderboard/arc_challeneg_25shots.yaml


recipes/evaluation/open_llm_leaderboard/hellaswag_10shots.yaml → tools/benchmarks/llm_eval_harness/open_llm_leaderboard/hellaswag_10shots.yaml


recipes/evaluation/open_llm_leaderboard/hellaswag_utils.py → tools/benchmarks/llm_eval_harness/open_llm_leaderboard/hellaswag_utils.py


recipes/evaluation/open_llm_leaderboard/mmlu_5shots.yaml → tools/benchmarks/llm_eval_harness/open_llm_leaderboard/mmlu_5shots.yaml


recipes/evaluation/open_llm_leaderboard/winogrande_5shots.yaml → tools/benchmarks/llm_eval_harness/open_llm_leaderboard/winogrande_5shots.yaml