Lead AI Test Automation Engineer

Reposted 3 Days Ago
Be an Early Applicant
Hiring Remotely in Philippines
Remote
Expert/Leader
Software
The Role
Seeking a Lead Test Automation Specialist to architect automation frameworks using Playwright, implement AI-driven testing, and enhance QA processes. Requires strategic thinking and hands-on execution to ensure software quality.
Summary Generated by Built In

About DevRev

At DevRev, we're building the future of work with Computer – your AI teammate. Unlike traditional tools, Computer unifies all your data sources, tools, and workflows into a single AI-ready platform, giving employees real-time insights, proactive suggestions, and powerful agentic actions. It extends your existing software with AI-native apps and agents that work alongside your teams and customers – updating workflows, coordinating across teams, and eliminating repetitive work. We call this Team Intelligence: human-AI collaboration that breaks down silos, brings people back together, and frees you to solve bigger problems. Backed by Khosla Ventures and Mayfield with $150M+ raised, DevRev is trusted by global companies across industries.

About the role:
As our Lead AI Test Automation Specialist, you'll develop testing strategies, evaluation frameworks, and quality metrics specifically designed for LLM-powered applications. This role requires a unique blend of QA expertise, understanding of GenAI behaviour, and automation skills to ensure our AI features are reliable, accurate, and trustworthy.


Key Responsibilities:

    • Design and implement comprehensive testing strategies for GenAI features, including conversational AI, agentic systems, and LLM-powered workflows
    • Develop automated test suites for prompt testing, including regression tests that detect unintended changes in model behaviour
    • Create evaluation frameworks to measure GenAI quality across multiple dimensions (accuracy, relevance, safety, consistency, latency)
    • Build and maintain test datasets and golden examples that represent diverse user scenarios and edge cases
    • Implement monitoring and alerting systems to detect quality degradation in production GenAI features
    • Perform adversarial testing to identify potential failures, hallucinations, biases, or security vulnerabilities in AI systems
    • Collaborate with engineers to define acceptance criteria and quality gates for AI feature releases
    • Develop tools and frameworks that make it easy for engineers to test their GenAI implementations
    • Conduct user acceptance testing and gather feedback on AI feature performance from internal users
    • Document testing procedures, known issues, and quality metrics in clear, accessible formats
    • Partner with Product and Design teams to ensure AI features meet user experience standards
    • Stay current with GenAI testing methodologies, tools, and industry best practices
    Your Qualifications
    • PRE or test engineering experience, preferably with AI/ML systems.
    • Strong understanding of GenAI technologies including LLMs, prompt engineering, and AI application patterns
    • Experience with test automation frameworks and scripting (Python, JavaScript, Selenium, Pytest)
    • Knowledge of software testing methodologies (functional, integration, regression, performance, security testing)
    • Ability to design test cases and evaluation criteria for non-deterministic systems
    • Strong analytical and problem-solving skills with attention to detail
    • Experience with API testing tools (Postman, REST Assured) and backend testing
    • Familiarity with CI/CD pipelines and automated testing integration
    • Excellent communication skills for documenting issues and collaboration
    Preferred Qualifications
    • Experience testing conversational AI, chatbots, or agentic systems
    • Knowledge of ML model evaluation metrics and techniques
    • Familiarity with LLM evaluation frameworks (LangSmith, PromptFoo, Ragas)
    • Experience with performance testing and load testing AI APIs
    • Understanding of responsible AI principles, including fairness, transparency, and safety testing
    • Background in enterprise software or SaaS QA
    • Experience with test management tools (TestRail, Zephyr, Jira)
    • Knowledge of security testing methodologies for AI systems
    • Scripting experience with Python, including working with LLM APIs
    What Makes This Role Exciting
    • Define Quality practices for GenAI applications
    • Work on cutting-edge AI technologies and help ensure they're reliable and trustworthy
    • Shape quality standards that will impact millions of enterprise users
    • Collaborate closely with engineers, data scientists, and product teams
    • Grow expertise in a highly specialized and increasingly important domain
    • Influence the entire AI product development lifecycle from design to release
    • Join a team that values quality as a first-class concern, not an afterthought

Join us in innovating our testing processes and ensuring the delivery of high-quality software products through advanced automation techniques.

DevRev is an equal opportunity employer and does not discriminate on the basis of race, gender, sexual orientation, gender identity/expression, national origin, disability, age, genetic information, veteran status, marital status, pregnancy or related condition, or any other basis protected by law.

Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
HQ: Palo Alto, CA
127 Employees
Year Founded: 2020

What We Do

DevRev is a business software company that brings developers (dev) and customers (rev) together in the era of product-led growth. The company is building an API-first dev-centric CRM that leverages data, design, and machine intelligence to empower devs to build, support, and grow their revs. In times of anemic NPS and high customer churn, DevRev strives to create the world's most customer-centric companies led by happy developers

Similar Jobs

Mondelēz International Logo Mondelēz International

S4 MTI Process Excellence & Value Realization Lead

Big Data • Food • Hardware • Machine Learning • Retail • Automation • Manufacturing
Remote or Hybrid
5 Locations
90000 Employees

Smartly Logo Smartly

Accountant

AdTech • Artificial Intelligence • Digital Media • Marketing Tech • Social Media • Software • Generative AI
Easy Apply
Remote or Hybrid
Philippines
805 Employees

Smartly Logo Smartly

Designer

AdTech • Artificial Intelligence • Digital Media • Marketing Tech • Social Media • Software • Generative AI
Easy Apply
Remote or Hybrid
Philippines
805 Employees
5-5 Annually

Optum Logo Optum

Healthcare Collections Associate - Quezon City

Artificial Intelligence • Big Data • Healthtech • Information Technology • Machine Learning • Software • Analytics
Remote
Metro Manila, PHL
160000 Employees

Similar Companies Hiring

Milestone Systems Thumbnail
Artificial Intelligence • Other • Security • Software • Analytics • Big Data Analytics
Lake Oswego, OR
1500 Employees
Fairly Even Thumbnail
Hardware • Other • Robotics • Sales • Software • Hospitality
New York, NY
30 Employees
Kepler  Thumbnail
Fintech • Software
New York, New York
6 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account