HomeCompaniesCumulus Labs
Cumulus Labs

Performant serverless GPU inference

Cumulus Labs is a serverless GPU cloud with a proprietary inference stack, purpose-built for AI teams who want faster performance, lower costs, and zero infrastructure work. If your team uses GPUs — for inference, training, fine-tuning, or any compute workload — Cumulus replaces the painful parts of your stack with a platform that just works. Most teams today are stuck choosing between bad options. Self-hosting inference means wrestling with vLLM or SGLang configurations, debugging CUDA issues, and babysitting infrastructure that breaks at scale. Managed API providers like OpenRouter or Fireworks are convenient but expensive, and you're paying their margin on top of the compute. Serverless GPU platforms give you flexibility but hit you with slow cold starts, idle billing, and no help on the inference layer — you're still on your own to make models run fast. Cumulus eliminates that tradeoff. The platform assigns containers in seconds, scales to zero when you're idle, and scales up to as many instances as you need with no waitlists or approvals. You're billed by the second for exactly the compute you use — nothing more. For inference, Cumulus ships Ion — a proprietary engine that supports all major LLMs, VLMs, and MoE architectures out of the box. Ion is optimized for latency and throughput beyond what teams typically achieve managing vLLM or SGLang themselves, with zero configuration required. Whether you're serving your own fine-tuned model or hosting an open-source model like Llama or Mixtral, Ion handles the performance layer so your team doesn't have to. Cumulus also supports checkpointing, model compilation, and LoRA serving, and the team forward-deploys custom optimizations directly for customers. For training, fine-tuning, and general container workloads, teams bring any job to Cumulus and run it on the same serverless infrastructure — no cluster management, no GPU debugging, no orchestration setup. If you're paying for inference APIs, Cumulus lets you run the same models yourself for less. If you're self-hosting, Cumulus makes your models faster without the operational burden. If you need GPUs for any workload at all, Cumulus is the simplest and most performant way to get them.
Active Founders
Suryaa Rajinikanth
Suryaa Rajinikanth
Founder
Suryaa Rajinikanth studied computer science at Georgia Tech, where he concurrently worked at TensorDock as a Lead Engineer, building the first distributed GPU marketplace serving thousands of consumers and businesses. He went on to deploy critical AI systems and infrastructure in high-performance environments at Palantir.
Veer Shah
Veer Shah
Founder
Veer studied Computer Science at the University of Wisconsin—Madison, graduating in December 2025. During college, he worked at an aerospace startup where he led a Space Force SBIR contract for military satellite communications and contributed to several NASA SBIR programs, two of which were commercialized and are currently being flight tested in space. Before college, he captained his FIRST Robotics Team 5422: Stormgears, qualifying for Worlds all four years.
Company Launches
Cumulus Labs ☁️ | Supercharge Your Training & Inference
See original launch post

TL;DR

Cumulus is a performant GPU cloud that preemptively optimizes your training and inference workloads across our global supply of multi-tenant clusters. The result: we save you 50-70% through charging by physical resources used, provide faster inference with ultra-low cold starts, and ensure zero time spent debugging infrastructure.

Ask: If you’re training models or serving inference workloads, frustrated with GPU costs and performance, let us optimize your LLMs, LoRAs, vision models, and more.

Website: https://cumuluslabs.io
Docs: https://docs.cumuluslabs.io

Launch Video: https://www.youtube.com/watch?v=duQwV50GKXc

The Problem

AI teams are bleeding money and time on GPU infrastructure:

  • Massive waste: Teams pay for idle GPUs sitting at 30-40% utilization because scaling is unpredictable
  • Infrastructure hell: Engineers spend weeks configuring Kubernetes, debugging OOM errors, and managing failovers instead of building models
  • Cold start latency: Inference workloads take 10-30+ seconds to spin up, killing user experience
  • Vendor lock-in: Once you commit to a cloud provider, switching costs make it nearly impossible to optimize for price or performance
  • Skyrocketing costs: Companies burn through runway 2-3x faster than planned because GPU bills spiral out of control

Every hour debugging infrastructure is an hour not spent improving your models. Every dollar wasted on idle GPUs is a dollar not spent on growth.


The Solution

Cumulus is a GPU optimization layer that makes compute cheap, fast, and invisible.

We aggregate compute from everywhere—big cloud providers, trusted data centers, individual hosts—into a single unified pool. Then we do three things no one else does:

1. Predictive Packing & Live Migration (Training/Fine-tuning)

Your training jobs are intelligently packed alongside other workloads to maximize GPU utilization. As your job runs, we predict resource usage and live-migrate you to faster or cheaper clusters without interruption. No more paying for an entire H100 when you only need 40% of it.

2. Execution State Capture & Global CDN (Inference)

We capture your model's live execution state (VRAM, memory, loaded weights) and replicate it across our global compute CDN. When a request comes in, we serve from the closest cluster with ultra fast cold starts—no more waiting 30 seconds for a job to spin up. We have tested with LLMs, vision models, LORAs, and many others.

3. Intelligent Scheduling & Auto-Recovery

Our scheduler constantly monitors your jobs, diagnoses failures, and auto-recovers without manual intervention. The Cumulus prediction system learns your usage patterns over time and pre-allocates resources before you need them.

The bottom line: You write 20 lines of config. We handle everything else.

Our Demo: https://www.youtube.com/watch?v=J0KRFWE3-fg

The Team

uploaded image

(winning the science fair in 4th grade for a robot that solved Rubik’s cubes)

Veer Shah (Founder)
Led a Space Force program and worked on ML workloads at an aerospace startup supporting NASA missions, where infrastructure needed to be both performant and secure.

Suryaa Rajinikanth (Founder)
Built custom GPU compute solutions at TensorDock, then moved to Palantir where he built critical infrastructure for the US Government. Deep expertise in distributed systems and resource optimization.

We met as third graders and have been building together our whole lives. We've seen the GPU infrastructure problem from both sides: Suryaa from the provider side at TensorDock, Veer from the customer side running mission-critical ML workloads. We started Cumulus Labs because we knew exactly what the industry needed—and no one was building it.

What We're Looking For

If you're training models or serving inference workloads, frustrated with vendor lock-in, or simply paying too much for your GPUs, reach out.

Know AI/ML teams experience any of these issues? Connect them with us.

We optimize LLMs, LoRAs, vision models, and more.

We'd love your feedback on which features matter most.


Get In Touch

Join the waitlist: https://cumuluslabs.io
Contact us: founders@cumuluslabs.io
Book a demo: Here










Huge thanks to partners, our batch-mates, and everyone who's helped so far.

Let's make AI infrastructure invisible.

— Veer, Suryaa, and the Cumulus team

uploaded image

Cumulus Labs
Founded:2025
Batch:Winter 2026
Team Size:2
Status:
Active
Location:San Francisco
Primary Partner:Jon Xu