MLOps Engineer (AI Platform)

Posted 4 Days Ago
Be an Early Applicant
5 Locations
In-Office or Remote
Senior level
Artificial Intelligence • Conversational AI
The Role
As an MLOps Engineer, you will architect and build scalable infrastructure for AI models, ensuring reliability and automation throughout the entire lifecycle of production AI systems.
Summary Generated by Built In

Are you ready to move beyond maintaining legacy systems and build something truly new? What if your next role gave you the keys to architect an entire AI platform from the ground up, powering systems that serve millions of users? We're looking for a foundational MLOps leader to design, build, and own the infrastructure that will define the future of our AI capabilities.

You will architect and build the automated, scalable infrastructure that powers our entire suite of AI models—from Agentic AI and NLU to Voice Biometrics and ASR—ensuring they operate flawlessly and securely for millions of users. You will make key technical decisions, establishing the patterns and practices that will guide our machine learning operations for years to come.

This is a hands-on technical leadership role for an engineer who wants to make a career-defining impact. You won't just be joining a team; you'll be setting the standard for how we build, deploy, and operate machine learning at scale.

Your Mission: Architecting the Future of AI Operations

As our first dedicated MLOps Architect, you will have the autonomy and resources to build our ML platform from scratch. You will be responsible for the entire lifecycle of our production AI systems, ensuring they are reliable, secure, and automated. You will:

  • Architect and build the automated, scalable infrastructure that powers our entire suite of AI models—from Agentic AI and NLU to Voice Biometrics and ASR—ensuring they operate flawlessly and securely for millions of users.
  • Make key technical decisions, establishing the patterns, tools, and best practices that will guide our machine learning operations for years to come.
  • Collaborate closely with world-class researchers, data scientists, ML engineers, and cloud architects to translate cutting-edge research into robust, production-grade products.
  • Champion a culture of automation, governance, and performance across all our AI/ML initiatives.

A Glimpse into Your Impact: Key Challenges You'll Own

  • Infrastructure as Code (IaC) Foundation: You will design and implement our entire MLOps infrastructure on AWS from the ground up using Terraform, establishing best practices for security, scalability, and cost-efficiency.
  • CI/CD for Machine Learning: You will build and own the end-to-end CI/CD pipelines using GitLab and Jenkins, automating everything from model training and validation to canary deployments and production rollbacks.
  • Containerization & Orchestration at Scale: You will lead the productization of our complex ML models, containerizing them with Docker and deploying them on a robust Kubernetes platform that you will help architect, build, and manage with Helm.
  • Proactive Observability: You will establish a culture of deep system insight by implementing and managing a comprehensive observability stack (e.g., Prometheus and Grafana), ensuring our models meet stringent performance, reliability, and security SLAs.

What You'll Bring to the Team

We are looking for an experienced engineer with a builder's mindset and a passion for creating elegant, scalable systems. You have a proven track record of operating critical infrastructure at scale and thrive in an environment where you can take ownership and drive technical strategy.


Requirements
  • 5+ years in a Senior DevOps, SRE, or MLOps role with a focus on production systems
  • Deep expertise in architecting and managing Kubernetes clusters in a production environment.
  • Proven mastery of at least one major IaC tool (Terraform is strongly preferred).
  • Strong proficiency in a systems-level scripting language (e.g., Python, Go).
  • A track record of building and maintaining CI/CD pipelines for critical production services.
  • Direct experience deploying and managing specific ML models (e.g., Agentic AI, NLU, ASR, TTS).
  • Experience with dedicated ML workflow orchestration tools (e.g., Kubeflow, Apache Airflow).
  • Familiarity with ML experiment tracking and model registry tools (e.g., MLflow, SageMaker Model Registry).
  • Experience deploying models on specialized hardware (e.g., GPUs, Inferentia, Trainium, etc.).

Benefits
  • Fixed compensation;
  • Long-term employment with the working days vacation;
  • Development in professional growth (courses, training, etc);
  • Being part of successful cutting-edge technology products that are making a global impact in the service industry;
  • Proficient and fun-to-work-with colleagues;
  • Apple gear.

Omilia is proud to be an equal opportunity employer and is dedicated to fostering a diverse and inclusive workplace. We believe that embracing diversity in all its forms enriches our workplace and drives our collective success. We are committed to creating an environment where everyone feels welcomed, valued, and empowered to contribute their unique perspectives without regard to factors such as race, color, religion, gender, gender identity or expression, sexual orientation, national origin, heredity, disability, age, or veteran status, all eligible candidates will be given consideration for employment.

Top Skills

AWS
Docker
Gitlab
Grafana
Helm
Jenkins
Kubernetes
Prometheus
Terraform
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
354 Employees
Year Founded: 2002

What We Do

At Omilia we are engaged to provide the most human-like human-to-machine communication experiences and technologies in order to help large enterprises improve the customer care experience.

Starting out of a small garage, Omilia is now serving 1 billion conversations, in 30 languages, across 17 countries.

With one of the fastest growing NLU solutions in the market, Omilia has been recognized as a Leader in the 2022 Gartner® Magic Quadrant™ for Enterprise Conversational AI Platforms, as well as in the IDC Marketscape for Worldwide Conversational AI Software Platforms for Customer Service 2021.

Our technology allows the enterprise to take advantage of Open-Question customer care with end-to-end Self-Service to greatly improve customer experience and significantly decrease operational costs.

In 2016 Omilia expanded to USA and Canada, counting 33 full production deployments worldwide and case studies with proven KPIs and ROIs across various industries.

Similar Jobs

Enverus Logo Enverus

Account Executive

Big Data • Information Technology • Software • Analytics • Energy
In-Office or Remote
2 Locations

Coupa Logo Coupa

Sales Executive

Artificial Intelligence • Fintech • Information Technology • Logistics • Payments • Business Intelligence • Generative AI
Remote
Spain

GitLab Logo GitLab

Back-end Engineer

Cloud • Security • Software • Cybersecurity • Automation
Easy Apply
Remote
31 Locations
98K-210K Annually

GitLab Logo GitLab

Back-end Engineer

Cloud • Security • Software • Cybersecurity • Automation
Easy Apply
Remote
31 Locations
98K-210K Annually

Similar Companies Hiring

Scrunch AI Thumbnail
Software • SEO • Marketing Tech • Information Technology • Artificial Intelligence
Salt Lake City, Utah
Credal.ai Thumbnail
Software • Security • Productivity • Machine Learning • Artificial Intelligence
Brooklyn, NY
Standard Template Labs Thumbnail
Software • Information Technology • Artificial Intelligence
New York, NY
10 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account