Viraat Das

Exla FLOPs - On-Demand GPU clusters - The Cheapest H100s Anywhere

Exla FLOPs is the only service where you can instantly spin up 64, 128, or more GPUs - no waitlists, no commitments. Just clusters at your command.

Add a comment

Replies

Best
Viraat Das

Hello!

I'm very excited to launch this product!

We built this because during our own AI training and fine‑tuning, we hit a wall when trying to scale past 8 GPUs. We were manually stitching together nodes across different clouds, so we decided to productize the solution.

Exla FLOPs has the lowest pricing of H100s among any cloud providers.

The cluster is built for developers with insane availabilities for all types of GPUs.

We’re thrilled to see what you build with it! Your feedback means a lot!

We are looking to give out free credits as well. Please fill this out and we'll be depositing credits soon after: https://tally.so/r/meGzgE

Roop Pal

@viraat_das This is solving such a desperate need as training is so expensive! Congrats on the launch and thank you for making this!

Viraat Das

@roop_pal Appreciate it Roop! Excited to see this being used to solve real problems!

Olga Shiryaeva

Congrats on the launch! This looks promising.

How do you handle node failures during long training runs? Any automatic recovery mechanisms?

What storage options do you provide for datasets and model checkpoints?

You mentioned "The cluster is built for developers with insane availabilities for all types of GPUs." - how do you get around chip shortages?

Viraat Das

@olga_s52 thank you!

  1. Node failures:

    Right now, Exla FLOPs gives you direct SSH access to your own bare metal nodes — no scheduler in the middle. This means you have full control over your setup (Slurm, DeepSpeed, Kubernetes, or otherwise). While we don’t currently handle automatic job recovery ourselves, most users bring their own checkpointing or orchestration setup to manage long training runs.

  2. Storage options:
    We provide fast, local NVMe on each node for high-speed I/O. For persistence across runs, users typically mount their own S3/GCS buckets or connect to remote storage solutions. Shared storage or NFS-style volumes are in the works.

  3. Chip availability:
    Rather than relying on a single cloud, we dynamically source idle GPU capacity across a network of bare metal providers, resale markets, and underutilized datacenter inventory. This allows us to provision 64+ GPUs seemlessly, even when traditional clouds are maxed out or heavily reserved.

Do reach out for any feedback or feature requests though: viraat@exla.ai

Prathamesh Ware

All the best for the launch.

Sawaram Suthar
All the best for launch.
Viraat Das

@sawarams Thank you!

Aarush Kukreja

curious to know how you're able to do this? sick product

Youssef Kallel

This is gonna be huge

Jesse Li

Cheaper than every major cloud provider is actually insane. Congrats!

Viraat Das

@jesseliii Thanks! Ez money!

Aaron Aftab

Great looking piece of software and even better looking CEO

Viraat Das

@aaronaftab takes a good looking CEO to make a good looking sfotware!

Kawin Nikomborirak

Incredible service at an incredible price point!

Viraat Das

@knikomborirak Thank you thank you!

Viraat Das

The form link has been corrected: https://tally.so/r/meGzgE