GPU Compute — Rent NVIDIA GPUs On Demand | VoltageGPU

GPU Pricing

  • NVIDIA RTX 4090 — $0.37/hr, 24 GB VRAM, best value for inference and fine-tuning
  • NVIDIA A100 — $2.02/hr, 80 GB VRAM, ideal for training and research
  • NVIDIA H100 — $2.77/hr, 80 GB VRAM, transformer engine with FP8 and NVLink
  • NVIDIA H200 — $4.07/hr, 141 GB HBM3e, large model training
  • NVIDIA B200 — $7.50/hr, 192 GB VRAM, frontier AI and FP8

Features

  • Per-second billing — pay only for what you use, no minimum commitment
  • SSH access — full root access to your GPU instance
  • Custom Docker templates — PyTorch, TensorFlow, Jupyter, Stable Diffusion pre-installed
  • Persistent volumes — NVMe storage that persists across pod restarts
  • Multi-GPU support — up to 8x H100 with NVLink for distributed training

Use Cases

  • AI training — train deep learning models, distributed training across multi-GPU nodes
  • Fine-tuning — LoRA, QLoRA, full fine-tuning of LLMs on consumer and enterprise GPUs
  • Inference — deploy models for real-time inference with vLLM, TGI, Triton
  • Rendering — 3D rendering, video encoding, VFX with Blender, Maya, Unreal Engine
  • Research — academic and scientific computing with GPU acceleration

How It Works

  1. Browse available GPUs — filter by model, VRAM, price, or availability
  2. Deploy your pod — select a template and launch in 30 seconds
  3. Connect via SSH — full root access with CUDA, drivers, and frameworks installed
  4. Pay per second — no minimum commitment, stop anytime
GPU Cloud Computing

Rent GPU Compute
From $0.22/hr

Deploy GPU instances in 30 seconds. Pre-installed CUDA, PyTorch, TensorFlow. From RTX 4090 to H200 — pay per second.

GPUs online
30s deploy time
$5 free credit
Per-second billing

Available GPUs

Prices update in real-time. Click any GPU to deploy instantly.

vs AWS
Up to 80%
cheaper on RTX 4090
vs RunPod
Up to 57%
cheaper on RTX 4090
vs Lambda
Up to 73%
cheaper on consumer GPUs
See full comparisons

Deploy in 3 Steps

From sign-up to running code in under a minute.

1
Choose a GPU
Pick from RTX 4090, A100, H100, H200 and more. Filter by VRAM, price, or availability.
2
Select a Template
PyTorch, TensorFlow, Jupyter, Stable Diffusion — pre-configured environments ready to go.
3
Connect & Run
SSH in or open Jupyter. Your GPU is ready with CUDA, drivers, and your framework installed.

Built for Every Workload

AI / ML Training
Train deep learning models, fine-tune LLMs, run distributed training across multi-GPU nodes.
PyTorchMulti-GPUNVLinkCheckpoints
ML Inference
Deploy models for real-time inference. vLLM, TGI, Triton — optimized for low latency.
vLLMTGIOpenAI APIAuto-scale
Rendering & Viz
3D rendering, video encoding, VFX. Blender, Maya, Unreal Engine with GPU acceleration.
BlenderOptiXRay TracingNVENC

FAQ

How much does it cost?

Starting at $0.22/hr with per-second billing. No minimum commitment. $5 free credit for new users.

How fast is deployment?

30-60 seconds. Select a GPU, pick a template, and you're connected via SSH or Jupyter instantly.

What software is pre-installed?

CUDA 12, cuDNN, PyTorch 2.x, TensorFlow, JAX, Transformers, Jupyter Lab, Docker — all ready to go.

Is data persistent?

Yes. All pods include persistent NVMe storage. Data is preserved when stopped. Volumes available for cross-pod storage.

What is Confidential Compute?

Encrypted GPU instances (H100, H200, B200) with hardware-level isolation. Your data stays private even from the infrastructure provider.

Can I use multiple GPUs?

Yes. Multi-GPU configurations up to 8x H100 with NVLink. Distributed training with torchrun and DeepSpeed supported.

Start Using GPU Compute

$5 free credit. No credit card required. Deploy in 30 seconds.

Per-second billingPre-installed ML stackBitcoin accepted24/7 support