
Hi YC, we’re Shreyas and Rohan — two ex-Berkeley students building Ashr.
TL;DR : Ashr catches bugs and validates agent response quality by mimicking the patterns and quirks of real users. Developers today struggle to simulate these conditions in tests, resulting in real-time fails and unexpected behavior. We integrate as an SDK within your codebase.
AI automations are top of mind for every developer: large companies are integrating AI into older systems to stay technologically relevant, while AI-native companies are quickly emerging in every vertical. Unfortunately, most of these AI tools miss the mark: constant hallucinations, unexpected behavior, and dubious quality have eroded trust in agentic systems. In most cases, this is not because foundation models aren’t powerful enough, but because of subtle prompting, tooling, and scaffolding errors that only emerge with rigorous testing. The need for high volumes of realistic testing arises in many scenarios: we’ve spoken to companies who want to prove that their agents are capable of handling sensitive data to potential customers, developers who want to make their agents work better, and engineers who want to test edge-cases and new features but don’t have all the necessary tools.
We’ve seen teams confront this problem with a litany of half-measures: pushing to production and tanking user trust as a means of discovering bugs, half-baked test harnesses, and vibe-coded test suites that only scratch the surface. We built ashr to solve this.
Ashr is a modular testing and evaluation platform for your agentic system. We integrate as an SDK into your codebase, allowing you to request test suites and analyzing your agents’ output based off your tooling and ideal outputs. These tests mimic how real users utilize an agentic system: taking complex tool call paths, using obscure language with unclear instructions, and attempting to use restricted features. We develop environments, fine-tuned custom scorers, and stop at nothing to provide all the insights needed to bring your agent to a truly production-ready state.
We also work seamlessly with Claude Code and Cursor, allowing your coding agents to quickly prototype and test changes while receiving immediate feedback. Teams have run swarms of agents testing with ashr overnight, and have woken up to changes that reduced hallucinations and increased accuracy and user retention, all without writing a single line of code.
If you’re building agents, book a call today at https://ashr.io/. All onboarding is done via call so we can ensure everyone is ready to go and start using as soon as possible. In addition, we’d love feedback on our docs, opinions about the nature of test infra, and just chatting with a bunch of builders on how to make our product better.
You can reach us at founders@ashr.io
We’re two young founders passionate about changing the world with agentic development