What are you trying to run?
Here are some common tools and the GPU specs they usually need:Tool | Recommended GPU | Notes |
---|---|---|
Fooocus | 1x RTX 4090 | Smoothest experience with 24GB VRAM |
Stable Diffusion | 1x RTX 3090 or better | Needs ~16GB VRAM+ for full models |
InvokeAI / AUTOMATIC1111 | 2x RTX 4090 | For faster batch processing |
ComfyUI | 1x RTX 4090 or higher | Works well with large workflows |
Choose the lowest configuration that runs your model comfortably. You can always scale up later.
What if you don’t know what to pick?
Start small and test.- Use 1x RTX 4090 to get going
- Monitor VRAM usage using
nvidia-smi
- If you hit limits or run into crashes, consider upgrading to 2x
How to change configuration later
You can’t resize an instance directly, but you can:- Stop or terminate the current instance
- Spin up a new one with a higher configuration
- Reinstall your app or transfer your data if needed
Save your model/data files before terminating an instance!
Want to go faster?
More GPUs = more parallelism. Use 2x configs if you:- Run large prompts or batch generations
- Train or fine-tune models
- Need lower latency for interactive work