Responsibilities
- Own an LLM vertical with a focus on a specific safety domain, technique, or use case (either from defense or red-team attack perspective)
- Generate high quality synthetic data, train LLMs, and conduct rigorous benchmarking.
- Deliver robust, scalable, and reproducible production code.
- Push the envelope by developing novel techniques and research that delivers the world’s most harmless and helpful models. Your research will directly empower our customers to more feasibly deploy safe and responsible LLMs.
- Co-author papers, patents, and presentations with our research team by integrating other members’ work with your vertical.
Qualifications
- Deep domain knowledge in LLM safety techniques.
- Extensive experience in designing, training, and implementing multiple different types of LLM models and architectures in the real world. Comfortability with leading end-to-end projects.
- Adaptability and flexibility. In both the academic and startup world, a new finding in the community may necessitate an abrupt shift in focus. You must be able to learn, implement, and extend state-of-the-art research.
- Preferred: past research or projects in either attacking or defending LLMs.
Top Skills
What We Do
Dynamo AI is pioneering the first end-to-end secure and compliant generative AI infrastructure that runs in any on-premise or cloud environment.
With a holistic approach to GenAI compliance, we help accelerate enterprise adoption to deploy secure, reliable, and compliant AI applications at scale.
Our platform includes three products:
- DynamoEval evaluates GenAI models for security, hallucination, privacy, and compliance risks.
- DynamoEnhance remediates identified risks, ensuring more reliable operations.
- DynamoGuard offers real-time guardrailing, customizable in natural language and with minimal latency
Our client base and partnerships include Fortune 1000 companies across all industries, which underscores our proven success in securing GenAI in highly regulated environments








