GPU Compute — Rent NVIDIA GPUs On Demand | VoltageGPU
GPU Pricing
- NVIDIA RTX 4090 — $0.37/hr, 24 GB VRAM, best value for inference and fine-tuning
- NVIDIA A100 — $2.02/hr, 80 GB VRAM, ideal for training and research
- NVIDIA H100 — $2.77/hr, 80 GB VRAM, transformer engine with FP8 and NVLink
- NVIDIA H200 — $4.07/hr, 141 GB HBM3e, large model training
- NVIDIA B200 — $7.50/hr, 192 GB VRAM, frontier AI and FP8
Features
- Per-second billing — pay only for what you use, no minimum commitment
- SSH access — full root access to your GPU instance
- Custom Docker templates — PyTorch, TensorFlow, Jupyter, Stable Diffusion pre-installed
- Persistent volumes — NVMe storage that persists across pod restarts
- Multi-GPU support — up to 8x H100 with NVLink for distributed training
Use Cases
- AI training — train deep learning models, distributed training across multi-GPU nodes
- Fine-tuning — LoRA, QLoRA, full fine-tuning of LLMs on consumer and enterprise GPUs
- Inference — deploy models for real-time inference with vLLM, TGI, Triton
- Rendering — 3D rendering, video encoding, VFX with Blender, Maya, Unreal Engine
- Research — academic and scientific computing with GPU acceleration
How It Works
- Browse available GPUs — filter by model, VRAM, price, or availability
- Deploy your pod — select a template and launch in 30 seconds
- Connect via SSH — full root access with CUDA, drivers, and frameworks installed
- Pay per second — no minimum commitment, stop anytime
GPU Cloud Computing
Rent GPU Compute
From $0.22/hr
Deploy GPU instances in 30 seconds. Pre-installed CUDA, PyTorch, TensorFlow. From RTX 4090 to H200 — pay per second.
Available GPUs
Prices update in real-time. Click any GPU to deploy instantly.
vs AWS
Up to 80%
cheaper on RTX 4090
vs RunPod
Up to 57%
cheaper on RTX 4090
vs Lambda
Up to 73%
cheaper on consumer GPUs
Deploy in 3 Steps
From sign-up to running code in under a minute.
1
Choose a GPU
Pick from RTX 4090, A100, H100, H200 and more. Filter by VRAM, price, or availability.
2
Select a Template
PyTorch, TensorFlow, Jupyter, Stable Diffusion — pre-configured environments ready to go.
3
Connect & Run
SSH in or open Jupyter. Your GPU is ready with CUDA, drivers, and your framework installed.
Built for Every Workload
AI / ML Training
Train deep learning models, fine-tune LLMs, run distributed training across multi-GPU nodes.
PyTorchMulti-GPUNVLinkCheckpoints
ML Inference
Deploy models for real-time inference. vLLM, TGI, Triton — optimized for low latency.
vLLMTGIOpenAI APIAuto-scale
Rendering & Viz
3D rendering, video encoding, VFX. Blender, Maya, Unreal Engine with GPU acceleration.
BlenderOptiXRay TracingNVENC
FAQ
How much does it cost?
Starting at $0.22/hr with per-second billing. No minimum commitment. $5 free credit for new users.
How fast is deployment?
30-60 seconds. Select a GPU, pick a template, and you're connected via SSH or Jupyter instantly.
What software is pre-installed?
CUDA 12, cuDNN, PyTorch 2.x, TensorFlow, JAX, Transformers, Jupyter Lab, Docker — all ready to go.
Is data persistent?
Yes. All pods include persistent NVMe storage. Data is preserved when stopped. Volumes available for cross-pod storage.
What is Confidential Compute?
Encrypted GPU instances (H100, H200, B200) with hardware-level isolation. Your data stays private even from the infrastructure provider.
Can I use multiple GPUs?
Yes. Multi-GPU configurations up to 8x H100 with NVLink. Distributed training with torchrun and DeepSpeed supported.
Start Using GPU Compute
$5 free credit. No credit card required. Deploy in 30 seconds.
Per-second billingPre-installed ML stackBitcoin accepted24/7 support