Expanded vLLM model catalog, faster launch times, and improved stability.New pre-packaged vLLM models added (total of 10 now available):
- Meta Llama-3.1 8B Instruct
- Mistral Small-3.1 24B Instruct
- Llama 3.3 70B Instruct
- Mistral Small-24B Instruct
- Qwen-2.5 VL 32B Instruct
- GPT OSS 20B
HTTPS support, new inference options, and smoother instance setup.HTTPS services are now available.vLLM inference servers added.Improved instance flow for smoother setup and management.More connectivity options introduced.
Instance controls, custom templates, and RTX 5090 support.Stop and start your Compute instances.Custom templates now supported.Added RTX 5090 support.