terraform.tfvars.example 666 B

123456789101112131415161718
  1. # Example terraform.tfvars for minimal Amazon SageMaker deployment
  2. # Copy this file to terraform.tfvars and customize as needed
  3. # AWS Configuration
  4. aws_region = "us-east-1"
  5. # Project Configuration
  6. project_name = "my-llama-api"
  7. environment = "dev"
  8. # Model Configuration
  9. model_image_uri = "763104351884.dkr.ecr.us-east-1.amazonaws.com/huggingface-pytorch-inference:2.0.0-transformers4.28.1-gpu-py310-cu118-ubuntu20.04"
  10. model_data_s3_path = "s3://my-bucket/model/model.tar.gz" # Update with your model path
  11. model_name = "Llama-3.3-70B-Instruct"
  12. # Instance Configuration
  13. instance_type = "ml.p4d.24xlarge" # GPU instance for Llama models
  14. initial_instance_count = 1