Home
Companies
Hatchet

Hatchet

Background task orchestration and visibility

Hatchet abstracts away the infrastructure for managing task queues and message brokers, so you can focus on writing your background task and workflow logic. With Hatchet, you can run slow OpenAI requests in the background with async tasks, chain complex tasks together into workflows, and set retries and timeouts to recover from failure.

Jobs at Hatchet

New York, NY, US
$150 - $210
0.50% - 2.00%
3+ years
Hatchet
Founded:2023
Team Size:2
Location:New York
Group Partner:Dalton Caldwell

Active Founders

Alexander Belanger

Co-founder & CEO at Hatchet (W24), previously co-founder & CTO at Porter (S20). M.S. in Physics, B.A. in Physics and Computer Science from the University of Pennsylvania. Always happy to chat anything devtools, OSS, infrastructure!

Alexander Belanger
Alexander Belanger
Hatchet

Gabe Ruttner

Co-founder & CTO at Hatchet (W24), previously co-founder & CTO at ClearMix (S20). Masters in CS/AI from Cornell. Always game to chat anything dev tools, OSS, AI Architecture!

Gabe Ruttner
Gabe Ruttner
Hatchet

Company Launches

tl;dr - At a certain scale, all applications need a task queue - think Celery for Python, BullMQ for Node.js, and Sidekiq for Rails. Hatchet is an open-source, managed task queue focused on scale, reliability, and improved developer experience.

https://youtu.be/3myPYVyaqwM

The problem

Building a reliable application is difficult — network connections can fail or be interrupted, the application could run out of memory, or a bad commit on production can take down your entire service. To reduce the pressure on APIs, task queues are typically used to offload different parts of the application logic to other services in the system.

However, these queues introduce more points of failure — they require new infrastructure, a new set of integrations to monitor and debug your tasks, and are difficult to use in development. As applications begin to scale, application developers are forced to implement solutions to avoid queue overflow and overcrowding and to think about how to optimize their queues.

With LLM APIs — which are high latency and highly variable in response quality — these problems are even worse. It’s now typical to ingest, sync, vectorize, and index the entirety of multiple data sources on account creation, and as LLM systems get more agentic, developers need to chain together unreliable LLM calls to fulfill requests. These requests are often slow and without the right architecture delivering a responsive and reliable user experience is challenging.

The solution

Hatchet is an end-to-end platform for task orchestration. It includes a web UI for monitoring tasks, an API for interacting with tasks and workflows, and SDKs for Python, Typescript, and Golang.

Hatchet has all the features you’d expect — like retries, timeouts, and cron schedules — along with features that developers are usually asked to build, like handling different fairness and concurrency patterns, durable execution, and result streaming.

The team

We’re both ex-CTOs and founders of multiple YC-funded companies — Alexander was co-founder and CTO at Porter (S20) while Gabe was co-founder and CTO at Clearmix (also S20). Before that, Alexander got his Master’s in Physics from UPenn while Gabe got his Master’s in AI from Cornell.

Alexander started working on this after years of wrangling task queues and distributed systems at both Porter and Oneleet, where he experienced the issues of building complex task orchestration and workflows. Gabe came at this from another perspective, having spent the last few years working with ML and more recently LLM stacks, where he saw the architectural problems that many of these applications have.

We immediately bonded over our excitement about what infrastructure will look like as AI becomes core to nearly every business.

How do I get started?

Hatchet is fully open source with an invite-only cloud offering. Feel free to request access or check out the self-hosting Hatchet docs.

If you’re dealing with observability problems in your queue, are running batch processing of documents for LLM workflows, or are experiencing queue fairness problems in multi-tenant apps, we’d love to chat.

And if you have any questions — or you’re just interested in distributed systems, task queues, and scheduling — we have a public Discord. See you there!