@@ -9,7 +9,7 @@ import sys
import fire
import yaml
-from llama_recipes.inference.model_utils import load_llama_from_config
+from llama_cookbook.inference.model_utils import load_llama_from_config
from transformers import AutoConfig, AutoTokenizer, MllamaProcessor
@@ -3,8 +3,8 @@
from warnings import warn
-from llama_recipes.configs import quantization_config as QUANT_CONFIG
-from llama_recipes.utils.config_utils import update_config
+from llama_cookbook.configs import quantization_config as QUANT_CONFIG
+from llama_cookbook.utils.config_utils import update_config
from peft import PeftModel
from transformers import (
AutoConfig,