The Role
Develop and run adversarial test suites for LLMs and image/video models, craft multilingual prompts, analyze outputs, and write vulnerability reports.
Summary Generated by Built In
About 10a Labs: 10a Labs is an applied research and AI security company trusted by AI unicorns, Fortune 10 companies, and U.S. tech leaders. We combine proprietary technology, deep expertise, and multilingual threat intelligence to detect abuse at scale. We also deliver state-of-the-art red teaming across high-impact security and safety challenges.
- Develop and run adversarial test suites—both manual and scripted—for LLMs and image / video models.
- Craft multilingual prompts, jailbreaks, and escalation chains targeting policy edge cases.
Analyze outputs, triage failures, and write concise vulnerability reports. - Contribute to internal tooling (e.g., prompt libraries, scenario generators, dashboards).
- Has 2-4 years of experience in red-teaming, security research, trust & safety, or related fields.
- Is comfortable scripting basic tests (Python, Bash, or similar) and working in Jupyter or prompt-engineering tools.
- Communicates clearly in English and at least one additional language (ideally major non-English language relevant to global threat landscapes).
- Thinks like an adversary, documents findings crisply, and iterates quickly.
- Bachelor’s degree—or equivalent experience—in CS, data science, linguistics, international studies, or security.
- Basic proficiency with Python and command-line tools.
- Demonstrated interest in AI safety, adversarial ML, or abuse detection.
- Strong writing skills for short vulnerability reports and long-form analyses.
- Ability to rapidly context switch across domains, modalities, and abuse areas
- Excited to work in a fast-paced and ambiguous space
- Full professional proficiency in Arabic, Chinese, Farsi, Portuguese, Russian, or Spanish, as well as English
- Prior work in content moderation, disinformation analysis, or cyber-threat intelligence.
- Experience with prompt-automation frameworks (e.g., Promptfoo, LangChain, Garak).
Familiarity with vector search or LLM fine-tuning workflows. - Formal training or certification in red-teaming or penetration testing.
- Salary range: $70K–$90K depending on experience.
- Opportunity for spot bonuses and annual performance-based bonus.
- Fully remote (U.S.-based) with flexible hours.
- Comprehensive health, dental, and vision.
- Generous PTO and paid holidays.
- 401(k) plan.
- Professional-development stipend for courses, conferences, or language study.
- We reward excellence with growth—team members who excel have clear paths for promotion and skill development.
Top Skills
Bash
Garak
Jupyter
Langchain
Prompt-Automation Frameworks
Prompt-Engineering Tools
Promptfoo
Python
Am I A Good Fit?
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.
Success! Refresh the page to see how your skills align with this role.
The Company
What We Do
10a Labs is an applied research and technology company specializing in AI security. We deliver intelligence collection, investigative research, and analysis for AI unicorns, Fortune 10 companies, and U.S. tech leaders.