diff --git a/llama/llama-3_1-8b-instruct/config.yaml b/llama/llama-3_1-8b-instruct/config.yaml index 755e3840..0425461e 100644 --- a/llama/llama-3_1-8b-instruct/config.yaml +++ b/llama/llama-3_1-8b-instruct/config.yaml @@ -1,7 +1,7 @@ model_name: "Llama 3.1 8B Instruct VLLM" python_version: py311 model_metadata: - example_model_input: {"prompt": "what is the meaning of life"} + example_model_input: {"prompt": "what is the meaning of life", "max_tokens": 1024} repo_id: meta-llama/Meta-Llama-3.1-8B-Instruct tensor_parallel: 1 requirements: