We want to build agents that autonomously validate code changes. Today that looks like AI that reviews pull requests in GitHub, catching bugs and enforcing standards. We’re reviewing close to 1B lines of code a month now for over 1,000 companies.
Problems we’re excited about
Coding standards can be idiosyncratic and are often poorly documented; can we build agents that learn them through osmosis like a new hire might?
Can we identify for each customer what types of PR feedback they do and don’t care about, perhaps using some sample efficient RL, in order to increase signal-to-noise ratio?
Some bugs are best caught by running the code, potentially against discerning AI-generated E2E tests. Can we autonomously deploy feature branches and use agents to parallel try to break the application to detect bugs?
Trajectory
Went from 0 ---> XM in <12 Months and growing >25% MoM
1,500+ customers
Raised 30M+ led by Benchmark, along with continued support from YC, Paul Graham, Initialized, SV Angel, etc..
Team
We have assembled a small, talent dense team who have scaled critical functions at companies like Stripe, Google, Figma, LinkedIn, etc.
Qualifications
BS in Computer Science or equivalent
Some research experience, ideally with ML/LMs/Agents
Strong programming skills and good product intuition
Responsibilities
Experiment with and apply recent advancements in agents/LMs etc. to improve the performance and capabilities of our products
Example: you might study multi-agent architectures, prototype and evaluate a multi-agent code review workflow, and then work with a team to integrate successful prototypes into production systems
Stay current with the latest research in LLMs, information retrieval, and developer tooling
Top Skills
What We Do
AI expert that understands your codebase, as an API.
Greptile can review your pull requests, answer questions about your codebase, write descriptions for JIRA tickets and more.








