COMPUTE

Vast.ai

Cheapest GPU rental in the market. Everything else is a trade-off. A peer-to-peer marketplace where individual hosts rent out spare hardware and users bid or buy on-demand — routinely 30–60% under RunPod Community on comparable GPUs, at the cost of operational polish.

RATING · 7.8 / 10 PRICING · MARKET-DRIVEN · RTX 3090 FROM $0.20/HR UPDATED · 2026-04-23
TRY VAST.AI → ESTIMATE MY SPEND → FAQ →

Estimate your monthly spend

INTERACTIVE · LIVE · MARKET-DRIVEN RATES

Vast.ai prices are market-driven — actual listings vary hour by hour, host by host. These are typical on-demand rates. Spot (interruptible) bids can be significantly lower. 720 hours is 24/7. Storage and bandwidth not included.

ESTIMATED MONTHLY SPEND
$40
USD / MONTH

On-demand rate · actual market listings will vary.

START ON VAST.AI →

BEST FOR

Research compute, training runs on a budget, sweep jobs, anyone trading polish for raw price.

NOT FOR

Production, compliance-sensitive workloads, zero-ops operation, enterprise procurement.

PRICING

Market-driven. RTX 3090 ~$0.20/hr · RTX 4090 ~$0.40/hr · A100 80GB ~$0.99/hr · H100 80GB ~$1.90/hr. Storage ~$0.10/GB/mo.

ALTERNATIVES

RunPod (more polish, similar model), Modal (Python-native, higher floor), Lambda Labs (bigger SLA), direct bare-metal.

What it is

Vast.ai is a peer-to-peer GPU marketplace. Individual hosts — datacenter operators, crypto miners repurposing hardware, academic clusters with spare capacity, a surprising number of hobbyists with rigs in closets — sign up to rent out their machines. Renters browse the resulting listings, filter by specs and reliability, and either buy on-demand or place a spot bid on interruptible capacity. The platform takes a cut, the host gets the rest, nobody has to sign a contract.

The company has been around since 2018, making it one of the older independent GPU marketplaces still operating. That tenure shows in the API, the CLI, and the filtering tools, which are more capable than first-time visitors expect. The dashboard itself is aggressively utilitarian — think eBay for GPUs — and the learning curve is steeper than a polished product like RunPod. The upside is the pricing: on almost any given GPU, the lowest listing on Vast.ai will be the lowest hourly rate you can find anywhere without signing a multi-year reserved-capacity deal.

The mental model that helps: Vast.ai is a marketplace, not a cloud. You're not renting from a company with a consistent operational standard; you're renting from whichever individual host happens to be listing the machine you want. That host might be a professional datacenter with 24/7 monitoring, or it might be a college student running a mining rig in a spare bedroom. Vast.ai's reliability score system — a numerical rating per host based on historical uptime, job-completion rate, and renter feedback — is how you tell them apart. Filter by reliability aggressively and the platform behaves more like a cloud; ignore it and you'll learn why the system exists.

The product surface breaks down into three main layers. On-demand instances are standard hourly rentals — you pick a listing, launch a Docker container or Jupyter notebook, pay by the second while it runs. Interruptible instances are spot-priced: you place a maximum bid, the host runs you when their capacity isn't sold at a higher price, and your job can be preempted at any moment if a higher bidder appears. Bulk ops via the CLI and API — scripting launches, teardowns, and image pushes — are how most serious users actually operate once past the first few runs.

Vast.ai's positioning against competitors is simple: cheaper. On 3090s, 4090s, A100s, and increasingly H100s, the floor price on Vast.ai consistently undercuts RunPod Community, which itself undercuts the hyperscalers by a wide margin. What you give up in return is polish, consistency, and — when things go wrong — someone to call.

What we tested

Across client engagements and internal experiments, we've put Vast.ai through the workloads it's actually designed for. We've run research training jobs on 3090s and 4090s for weeks at a time, filtered aggressively by reliability score. We've batch-processed large datasets on A100 80GB listings where the per-hour savings over RunPod Community made a real dent in total project cost. We've stood up multi-node sweep jobs via the CLI — launching ten interruptible instances in parallel to hyperparameter-search a model — and watched them run to completion. We've tested the spot-bidding model on off-peak 3090 listings to see what actually happens when a higher bidder shows up.

Hardware coverage: RTX 3080, 3090, 4090, A4000, A5000, A6000, A40, A100 40GB and 80GB (both PCIe and SXM variants where available), H100 PCIe and SXM, and a rotating menu of older and consumer SKUs (V100, P100, Titans, even 2080 Tis on the budget end). Some of the most interesting price-performance comes from GPUs the hyperscalers don't even offer — nobody rents you a 4090 on AWS, and the 4090 is an astonishingly capable card for a specific class of workload.

Operationally, we evaluated five things. First, raw price-performance: on a matched-GPU basis, how far under the next-cheapest competitor does Vast.ai sit? Second, host reliability variance: how different is a 0.99-score host from a 0.85-score host in real operational terms? Third, spot interruption behavior: when a higher bidder takes your capacity, how much warning do you get, and is your data recoverable? Fourth, CLI and API ergonomics for bulk ops: can you actually run a sweep job here, or does the tooling get in the way? Fifth, the shape of failure modes: when things break, what breaks, and what does the recovery path look like?

We also ran a direct head-to-head against RunPod Community on two matched workloads — a 13B LoRA fine-tune and a 50K-item batch inference run — to see how much the price delta survived contact with the operational overhead. None of what follows is a formal benchmark. What we can offer is the texture of actually running research-grade workloads on Vast.ai in 2025–2026 and living with the results.

Pricing, in detail

TYPICAL ON-DEMAND LISTINGS · 2026-04
RTX 3090 · 24GB
$0.20/ HR

The budget king. Floor listings dip to $0.12/hr. Fantastic for 7B inference and research prototyping.

  • 24GB VRAM, Ampere architecture
  • Per-second billing on most hosts
  • Often cheaper than a coffee for a full day's work
RTX 4090 · 24GB
$0.40/ HR

Consumer flagship. Listings start around $0.29/hr, stable on-demand $0.39–$0.59. Hyperscalers don't offer this card.

  • Ada Lovelace, FP8 support
  • Killer for 13B inference and LoRA
  • Roughly half the price of RunPod's 4090
H100 · 80GB
$1.90/ HR

Floor listings around $1.47/hr PCIe. Consistently the cheapest H100 listings in the market.

  • FP8 + Transformer Engine gains
  • 70B fine-tuning with multi-GPU
  • Availability tightens at peak hours
OLDER SKUs · V100, P100, A4000
FROM $0.08/ HR

Previous-generation cards at fractions of modern prices. Great for dev loops and lightweight inference.

  • V100 16GB, P100 16GB, A4000 16GB
  • Perfect for sanity-checking code before a real run
  • Rotating inventory — listings come and go
SPOT / INTERRUPTIBLE
FROM $0.10/ HR

Bid below on-demand, get preempted when a higher bidder arrives. Off-peak 3090 bids clear at absurdly low rates.

  • You set the maximum price you'll pay
  • Checkpoint or lose progress on preemption
  • Best used for sweep jobs, not single long runs
STORAGE · HOST-DEPENDENT

Storage is priced per host, typically $0.10/GB/mo on running instances. Not persistent by default — if the instance is destroyed, your volume goes with it. Plan for external storage (S3, rsync, HF datasets) from day one.

BROWSE LISTINGS →
SPOT BIDDING · READ THIS FIRST

Interruptible instances let you name a maximum price below on-demand. Off-peak (weekend, overnight) you'll often clear at the floor — a 3090 for $0.10/hr is realistic, an H100 under $1/hr happens. Preemption comes with a short grace period. Design your jobs to checkpoint every few minutes and this tier is a research budget multiplier. Ignore checkpoints and you'll pay the difference in lost training time the first time a higher bidder shows up at 3am.

What's good

The single biggest reason to use Vast.ai is that it is genuinely, consistently, the cheapest place to rent a modern GPU without signing a contract. That's not marketing — it's the output of a marketplace with enough supply that price competition actually happens. On the A100 80GB tier we routinely see listings 35–50% below RunPod Community, and RunPod Community is already the second-cheapest option most of the time. Over a long training run, that compounds into real money.

The bidding model is a legitimate research budget multiplier for anyone doing work that can be checkpointed. Hyperparameter sweeps, ablation runs, long-horizon training where you're already saving state every few epochs — all of this maps cleanly onto interruptible instances. Your worst case is "job restarts on a different host," which is a solved problem in modern ML frameworks. We've seen research projects cut their compute budget by 60% moving from on-demand to spot, with no meaningful difference in total wall-clock time.

Consumer GPUs that hyperscalers won't sell you are one of the most underrated features. You cannot rent a 3090 or a 4090 on AWS or GCP — it's not that they're expensive, they literally don't exist in those catalogs. For a lot of 7B–13B workloads, a 4090 at $0.40/hr outperforms an A10G at $1.20/hr on a hyperscaler. The price delta is ridiculous and the performance delta goes the wrong way for the hyperscaler. If your model fits in 24GB of VRAM, Vast.ai's consumer-card inventory is probably the best deal in AI infrastructure right now.

The CLI and API are surprisingly capable for bulk ops. Launching ten instances in parallel, mounting the same Docker image, running a sweep, and tearing it all down — it's maybe fifty lines of Python once you've learned the API. For anyone doing real research work, this is the "escape hatch" that turns Vast.ai from a weird marketplace into a first-class research tool. The dashboard is rough; the CLI is fine; the API is legitimately good.

The reliability score is a real filtering tool, not theater. Filter for hosts above 0.99 reliability, above a minimum number of completed jobs, with a history of not preempting on-demand, and the failure rate drops to "occasional" rather than "frequent." It's on you to do this filtering — the platform won't do it for you — but the data is exposed honestly enough that a savvy user can approximate RunPod-Community reliability at Vast.ai pricing.

Where Vast.ai earns its keep

Users report that Vast.ai feels like a marketplace operated by engineers who expect you to be an engineer too. The cost of that bargain is doing more of the operational work yourself; the benefit is paying less than anyone else on the planet for the exact same silicon.

One more small thing that matters: there is a real community of researchers, indie ML practitioners, and grad students who treat Vast.ai as their default compute layer. The Discord is active, the documentation has improved noticeably over the last two years, and the mental model — once you've run a job or two — transfers cleanly between users. This is a tool that rewards investment.

Pros & cons

OUR HONEST TAKE

WHAT WORKS

  • Lowest listed prices on almost every GPU tier — 30–60% under RunPod Community.
  • Spot bidding is a research-budget multiplier for checkpointable work.
  • Consumer cards (3090, 4090) hyperscalers don't even catalog.
  • CLI and API make parallel launch for sweep jobs a real workflow.
  • Reliability score is an actual signal when filtered aggressively.
  • Docker-first — local images run unchanged, no proprietary SDK.
  • Per-second billing on most hosts, no minimum commitment.

WHAT DOESN'T

  • Reliability varies dramatically by host — a 0.95 host and a 0.99 host are different products.
  • Storage is not persistent by default; destroying an instance loses the volume.
  • Dashboard UX is noticeably rougher than RunPod — utilitarian at best.
  • No enterprise compliance story: no SOC 2, no HIPAA, no signed BAAs.
  • Support is community-driven (Discord); no named contacts, no SLA.
  • Data-residency is whatever the host's country is — fine for research, wrong for regulated data.
  • SSH and Docker environments can differ between hosts in subtle ways that break scripts.

Common pitfalls

Six failure modes come up repeatedly across the teams we've seen move onto Vast.ai. None of them are dealbreakers, all of them benefit from naming and planning around.

Not filtering by reliability score. This is the single most common mistake first-time users make. The default listing view sorts by price, which surfaces the sketchiest hosts at the top. A $0.12/hr 3090 from a host with a 0.85 reliability score is not the same product as a $0.22/hr 3090 from a 0.99-score host — the first one will routinely disappear mid-job, while the second one behaves like infrastructure. Always set a reliability-score floor (we use 0.99 for production-adjacent work, 0.97 for research) before the first sort by price. The savings that remain are still enormous.

Expecting persistent storage. Vast.ai instance storage lives with the instance. Destroy the instance and the volume is gone. This catches people who have an AWS mental model where EBS survives EC2 destruction. Plan for external storage from day one: push model weights to Hugging Face or S3, stream datasets from external storage, and treat the local volume as purely ephemeral. Some hosts offer storage that survives stop/start, but not destroy — read the host's listing carefully.

Running production traffic on Vast.ai. This is the big one. Vast.ai is not a production platform. Hosts go offline for maintenance, reliability scores drift, spot instances get preempted, the marketplace nature means that "the exact host you're using" might not be available tomorrow. We have seen teams try to serve customer traffic from Vast.ai to save money, and the outage stories are consistently bad. Use it for training, batch inference, research experiments — not for anything with a user on the other end expecting an SLA.

Assuming SSH and Docker work identically across hosts. They don't, not quite. One host might have an older Docker daemon, a different CUDA base image in their default template, a non-standard SSH port, or network egress rules that break your download. Most of these are ten-minute problems once you've hit them, but they're real friction the first time. Script your environment setup idempotently — test that a fresh container reaches a known state — so host variance is absorbed by the script rather than by your workflow.

Bidding without understanding preemption risk. Spot is great — if your job can be preempted. If your model fine-tune takes 14 hours and your code doesn't checkpoint, preemption at hour 12 wastes the whole run. Before using spot, verify that your training loop checkpoints every N minutes, that the checkpoint actually restores, and that restart-from-checkpoint is part of your launch script. Test by killing your own instance once. If that's not how your code behaves, stick to on-demand until it is.

Sending regulated data anywhere near this platform. PHI, PII, financial records, customer data that your privacy policy commits to protecting — none of this belongs on a peer-to-peer marketplace where the host is a stranger with physical access to the machine. Vast.ai doesn't offer HIPAA BAAs, doesn't publish SOC 2 reports, and can't guarantee data residency in any specific jurisdiction. For those workloads, you need a compliance-posture cloud, full stop. For synthetic data, public datasets, model weights, and research-grade workloads, Vast.ai is fine. Know which you have.

What's actually offered

CAPABILITIES AT A GLANCE
PEER-TO-PEER MARKETPLACE

Hosts list machines, renters pick listings. Vast.ai matches, doesn't operate the hardware.

ON-DEMAND + INTERRUPTIBLE

Pay a stable hourly rate, or place a spot bid and get preempted when a higher bidder arrives.

HOST RELIABILITY SCORES

Numerical per-host rating based on historical uptime and job completion. Filter hard.

DOCKER + JUPYTER

Launch your own Docker image, or a ready-made Jupyter template. Pick whichever fits your loop.

CLI + API

Script launch, teardown, and image management. Bulk ops are what turn this into a real research tool.

SPOT BIDDING

Set a maximum you'll pay per hour; run when capacity is available below your bid. Huge savings for checkpointable work.

CONSUMER GPUs AVAILABLE

3090, 4090, Titan — GPUs hyperscalers don't offer, at prices that make the workload shape obvious.

BULK OPS + SWEEPS

Parallel launch for hyperparameter search, ablation runs, or any job that fans out across many instances.

SEEN ENOUGH?

Under $5 gets you a full day on a 3090. Nothing else in the market is priced like this.

TRY VAST.AI →

What's not

Let's be direct about the limits. The single biggest category of workload that does not belong on Vast.ai is anything with an end user on the other end of the connection expecting uptime. Production inference APIs, customer-facing demos, anything with an SLA in the contract — this isn't the product. The price comes from the marketplace structure, and the marketplace structure means no single operational standard, no capacity guarantee, no named account manager to call at 3am when something's wrong.

Compliance is the clearest limit. There is no SOC 2 Type II report you can hand to an enterprise client's security team. There is no HIPAA BAA. There is no real data-residency guarantee — your instance runs wherever your chosen host's datacenter happens to be, which might be a country you didn't intend to send data to. For anything that goes through a serious procurement review at a regulated organization, Vast.ai doesn't clear the bar. Not a criticism — the product isn't designed to clear that bar — but you need to know it.

Operational polish lags. The dashboard is functional but clearly not prioritized. Filtering is usable but not obvious. Billing is transparent but presented in a way that rewards familiarity rather than guiding new users. Error messages on failed launches are sometimes terse in a way that sends you to the Discord. Compared to RunPod, which feels like a polished product, Vast.ai feels like an engineer's tool that happens to have a web interface.

Support is community-driven. Discord is active, the Vast.ai team does respond there, and community members are often helpful — but there's no paid-tier support with named contacts, no 24/7 response guarantee, and no escalation path when a host disappears mid-job. For research, fine. For production, not fine.

Host variance on nominally identical hardware is a real thing. Two listings for "A100 80GB, 1 GPU, 16 vCPU" from two different hosts can have meaningfully different storage throughput, network egress performance, and even thermal behavior (a throttled GPU under sustained load produces different wall-clock times). If predictable performance matters to you, test on the specific host before committing to a long run.

Who should use it

Vast.ai is the right call if you fit one of four profiles.

The ML researcher or grad student on a budget, running experiments, ablations, and small training runs where the compute budget is coming out of a grant or a TA stipend. Vast.ai exists for this person. A 3090 at $0.20/hr means a 100-hour training run costs twenty dollars, which is tractable on any research budget. Add spot bidding and that number drops further. Compared to a university cluster with queue times measured in days, Vast.ai is a productivity multiplier.

The ML engineer running sweep jobs, fanning out across tens or hundreds of instances to explore a hyperparameter space. This is Vast.ai's sweet spot — bulk ops via the API, spot pricing, per-second billing, all stacking into a workflow that would cost 3–5× more anywhere else. A full ablation on 50 configurations across 8 seeds each might run you $200 on Vast.ai and $800 on RunPod Community. For academic publication timelines, this is the difference between "we ran the full sweep" and "we picked three settings."

The indie developer training on their own data, where the data isn't regulated, the model is for internal use or a public release, and the compute bill is a real line item in a personal budget. Fine-tuning a 7B model on a domain-specific corpus, training a small image model for a side project, experimenting with a new architecture — Vast.ai lets these projects happen at a fraction of the cost they'd otherwise need.

The team doing batch inference or dataset preprocessing at scale, where the job is embarrassingly parallel and the cost savings per-hour translate directly into project budget. Running 10 million items through a small classifier, generating embeddings for a large corpus, transcribing hundreds of hours of audio — these workloads map cleanly onto Vast.ai's economics.

Who should not use it: anyone running customer-facing production traffic, anyone in a regulated industry moving PHI or PII, anyone whose procurement requires SOC 2 or a signed BAA, and anyone whose team doesn't have the operational bandwidth to handle occasional host failures and spot preemption. For those cases, use RunPod Secure Cloud, Modal, Lambda Labs, or the hyperscaler premium — each of which buys something real that Vast.ai doesn't ship.

Compared specifically to the alternatives: RunPod is the next step up — more polish, better dashboard, still-aggressive pricing, worth the premium if reliability matters. Modal is a noticeably nicer Python-first experience with automatic cold-start handling and a higher price floor. Replicate is the fastest way to hit a specific open-source model as an API. Lambda Labs sits closer to a traditional cloud and is the safer pick for uptime-first workloads. None of them will be cheaper than Vast.ai on a matched GPU; all of them will be smoother in ways that matter for some users.

Verdict

Vast.ai is what happens when you let a GPU marketplace run at full price competition. The prices are real, the savings are real, and the operational trade-offs are real. For research, for sweep jobs, for indie ML work, for anyone who would rather spend an afternoon learning the CLI than spend 40% more for a polished dashboard — this is the right tool, and it has been the right tool for a while.

We rate it 7.8 / 10. It loses points for the operational surface, the missing compliance story, and the host variance that will always be part of the marketplace model. It gains them for pricing that beats everyone else, spot bidding that genuinely multiplies research budgets, and an API surface that makes bulk ops a first-class workflow. If we were rating purely on "dollars saved per training run," the number would be higher. If we were rating on "smoothness of the experience," the number would be lower. The 7.8 reflects that tension honestly.

If you're on the fence, filter for a 0.99-reliability 3090 and run a fine-tune for an afternoon. Two dollars later you'll know whether the platform fits your work. Most researchers discover it does. Most enterprise buyers discover it doesn't. The surprise is how many people move some of their workload here permanently once they've learned where the edges are.

Frequently asked

TAP TO EXPAND

Pick Vast.ai if price is the dominant factor, your workload is checkpointable, and you're comfortable doing more operational work yourself. Pick RunPod (especially Secure Cloud) if reliability matters more than the last 30% of savings, if you want a polished dashboard, or if you need persistent network volumes that survive instance destruction. Both are Docker-first and have active communities — the real difference is operational polish versus raw price.

Set a reliability score floor of 0.99 for any run longer than a few hours, and 0.97 for casual experimentation. Sort by the number of completed jobs — hosts with hundreds of completions behave differently than hosts with ten. Check for hosts with a "verified datacenter" badge where available. Read the host's storage and network notes; variance is real. The listings that remain after those filters are meaningfully more reliable than the raw cheapest listings at the top of the default sort.

No. This isn't a hedge — Vast.ai is a peer-to-peer marketplace where the host of your instance is a third party with physical access to the machine. There is no HIPAA BAA, no SOC 2 Type II report, and no data-residency guarantee. For PHI, PII, financial records, or any regulated data, use a hyperscaler with the appropriate compliance posture, or a compliance-forward GPU cloud with explicit BAAs. Vast.ai is fine for synthetic data, public datasets, open-weights model training, and research-grade workloads — know which you have before launching.

Three rules. First, your training loop must checkpoint every few minutes and reliably restore from the last checkpoint — test this by killing your own instance once before committing to a long run. Second, bid modestly below on-demand rates, not at the absolute floor; the floor bids get preempted first when demand rises. Third, use spot for sweeps, not for a single irreplaceable long run. A hyperparameter sweep of 50 configurations can absorb a few preemptions gracefully; a 14-hour fine-tune you're already behind on cannot.

Technically you can; we don't recommend it. Hosts go offline, reliability drifts, spot gets preempted, and when something breaks there is no named support contact. For production inference, we use either RunPod Secure Cloud, Modal, or a hyperscaler depending on the workload's sensitivity and scale. Vast.ai is where the development and training runs happen — the production artifact lands elsewhere.

Yes, and it's one of the best reasons to use the platform. Vast.ai has consistent inventory of RTX 3090s and 4090s — consumer GPUs that AWS, GCP, and Azure simply don't catalog. For 7B–13B inference and LoRA fine-tuning, a 4090 at $0.40/hr will outperform an A10G at $1.20/hr on a hyperscaler. If your workload fits in 24GB of VRAM, the 4090 inventory is arguably the best value in GPU rental anywhere.

Yes, with two caveats. First, pick a host with a high reliability score (0.99+) and ideally a verified-datacenter badge; the failure rate on well-filtered on-demand hosts is low enough for week-long runs. Second, checkpoint to external storage every hour or two — push to Hugging Face, S3, or rsync off the box. Instance-local storage is not persistent across destroy events. With those two in place, we've run multi-week training on Vast.ai without drama. Without them, expect to learn a lesson.

DONE READING?

Two dollars and an afternoon on a 3090 is the fastest way to know if the platform fits your work.

TRY VAST.AI → RE-RUN CALCULATOR →

[ INSTANT COMPARE ]

vs

Ready to rent a GPU for the price of a coffee? Takes about five minutes.

TRY VAST.AI → OR SCOPE A BUILD WITH US →