The Role
The Staff Software Engineer will build backend systems for LLM applications in healthcare, focusing on API design, data pipelines, observability, and system performance optimization.
Summary Generated by Built In
SuperDial is seeking a Staff Software Engineer, Applied AI to build and scale the backend systems that power LLM applications in healthcare. This role is ideal for an engineer who thrives at the intersection of backend architecture and applied AI, designing APIs, pipelines, and infrastructure that make LLMs reliable, secure, and cost-efficient in production. If you want to push LLMs beyond demos into mission-critical healthcare workflows, we’d love to hear from you.
About the Role:
- Backend for LLMs – Architect and implement scalable, low-latency APIs and services that wrap, orchestrate, and optimize LLMs for healthcare use cases.
- Data & Retrieval Pipelines – Build ingestion, preprocessing, and retrieval-augmented generation (RAG) pipelines to ground LLMs in clinical and revenue-cycle data.
- LLMOps & Observability – Design systems for model monitoring, evaluation, cost tracking, and guardrails, ensuring reliability and responsible use.
- Performance & Optimization – Engineer solutions for caching, batching, load balancing, and scaling LLM workloads across cloud and containerized environments.
- Security & Compliance – Implement HIPAA-ready infrastructure, data governance, and auditability for LLM-powered applications.
- Cross-Functional Collaboration – Partner with product, ML engineers, and healthcare experts to translate business workflows into robust backend systems.
- Technical Leadership – Drive end-to-end delivery of LLM backend projects, establish engineering best practices, and mentor peers in LLM system design.
About You:
- 5+ years of backend or full-stack software engineering experience, with 3+ years working on ML/LLM-enabled applications.
- Strong coding skills in Python (and ideally one statically typed language such as Go, Java, or TypeScript).
- Experience with LLM integration frameworks (Hugging Face, LangChain, LlamaIndex, OpenAI APIs, Anthropic, etc.).
- Deep knowledge of distributed systems, service-oriented architecture, and building APIs at scale.
- Cloud-native expertise: AWS/GCP/Azure, Kubernetes, Docker, Terraform, etc.
- Familiarity with MLOps/LLMOps practices: CI/CD for models, evaluation harnesses, monitoring, and reproducibility.
- Excellent system design skills and the ability to align technical architecture with product goals.
Preferred Qualifications:
- Experience applying LLMs in healthcare or other regulated industries (FHIR, HL7, HIPAA).
- Hands-on experience with RAG pipelines, vector databases, and structured-output orchestration.
- Background in enterprise SaaS or mission-critical platforms where uptime, latency, and scale matter.
- Knowledge of responsible AI, safety, and privacy-preserving ML techniques.
- The opportunity to apply cutting-edge AI to one of the world’s most important industries.
- A leadership role with ownership over core ML/LLM systems and influence on technical direction.
- Competitive salary, equity options, and benefits, including health, dental, and vision coverage.
The base pay range for this role is $200,000 – $275,000 per year.
Am I A Good Fit?
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.
Success! Refresh the page to see how your skills align with this role.
The Company
What We Do
OpenReq is a recruiting firm specializing in providing talent acquisition and staffing solutions for early-stage startups, particularly in the AI and Hard Tech sectors. They focus on technical and operations roles, helping founders manage the end-to-end recruiting process.







.png)