Deploy Mistral AI’s Voxtral on Amazon SageMaker AI
Configure your model in code/serving.properties: To deploy Voxtral-Mini, use the following code: option.model_id=mistralai/Voxtral-Mini-3B-2507 option.tensor_parallel_degree=1 To deploy Voxtral-Small, use the following code: option.model_id=mistralai/Voxtral-Small-24B-2507 option.tensor_parallel_degree=4 Open and run Voxtral-vLLM-BYOC-SageMaker.ipynb to deploy your endpoint and test with text, audio, and function calling capabilities. Docker container configuration The GitHub repo contains the full Dockerfile. The following code snippet highlights…


