- Train ML models via state of the art clusters
Our A100 distributed training clusters leverage a rail-optimized design using NVIDIA Quantum Infiniband networking and in-network collections using NVIDIA SHARP to deliver the highest distributed training performance possible. - Serve ML models with the fastest spin up times and responsive auto-scaling
We help you serve models as efficiently as possible with our proprietary auto-scaling technology and spin up times in as little as 5 seconds. Data centers across the country minimize latency delivering superior performance for all of your end users. - Directly access "bare metal" Kubernetes environments without the hassle of managing infrastructure
Our GPUs are accessible by deploying containerized workloads for increased portability, less complexity and overall lower costs. Not a Kubernetes expert? Our engineers are here to help.
Looking for a modern cloud, purpose-built for cutting edge AI?
CoreWeave empowers you to train, fine-tune, and serve models up to 35x faster while saving up to 80% on your current GPU costs.
Schedule a ChatUnparallelled performance for your most complex workloads
Not only do you get on-demand access to the industry’s broadest range of NVIDIA GPUs, but our Kubernetes-native infrastructure also delivers lightning quick spin-up times, responsive auto-scaling, and modern networking architecture.
Performance-Adjusted Cost Structure + Broadest Range of NVIDIA GPUs
- For on-demand workloads, we offer a la carte pricing (see below), where the total instance cost is a combination of a GPU component, the number of vCPU, and the amount of RAM allocated per hour.
- For workloads with relatively predictable usage patterns, we offer two types of volumetric discounts. Reserved Instances may be discounted up to 60% for 24/7 committed usage. Bulk Credits may be discounted up to 25% for larger-scale, burst compute use cases consumed on-demand.
- We DO NOT CHARGE for things like region-to-region transfers, workstation data, or egress in the vast majority of use cases.
GPU Model
VRAM (GB)
Max vCPUs per GPU
($0.01/hr)
($0.01/hr)
Max RAM (GB) per GPU
($0.005/hr)
($0.005/hr)
GPU Component
Cost Per Hour
Cost Per Hour
80 VRAM
48 Max CPUs
256 Max RAM
A100 80GB NVLINK
$2.21 / Hour
80 VRAM
48 Max CPUs
256 Max RAM
A100 80GB PCIe
$2.21 / Hour
80 VRAM
48 Max CPUs
256 Max RAM
A100 40GB NVLINK
$2.16 / Hour
40 VRAM
48 Max CPUs
256 Max RAM
A100 40GB PCIe
$2.06 / Hour
40 VRAM
48 Max CPUs
256 Max RAM
A40
$1.28 / Hour
48 VRAM
48 Max CPUs
256 Max RAM
RTX A6000
$1.28 / Hour
48 VRAM
48 Max CPUs
256 Max RAM
RTX A5000
$0.77 / Hour
24 VRAM
36 Max CPUs
128 Max RAM
RTX A4000
$0.61 / Hour
16 VRAM
36 Max CPUs
128 Max RAM
Quadro RTX 5000
$0.57 / Hour
16 VRAM
36 Max CPUs
128 Max RAM
Quadro RTX 4000
$0.24 / Hour
8 VRAM
36 Max CPUs
128 Max RAM
Tesla V100 NVLINK
$0.80 / Hour
16 VRAM
36 Max CPUs
128 Max RAM
What Others Say
From our clients to our partners, we strive to provide best-in-class solutions to
drive innovation and fast, flexible experiences.
drive innovation and fast, flexible experiences.
Ready to get started?
A member of our team will reach out within 24 hours... but probably sooner