Key Responsibilities
- Design, build, and maintain data pipelines and ETL/ELT processes to support AI and analytics initiatives across multiple state agencies.
- Develop and optimize data lakes and data warehouse architectures in cloud environments (e.g., AWS, Azure, GCP).
- Collaborate with data scientists, analysts, and application developers to operationalize AI models and analytical workflows.
- Implement data quality frameworks, governance policies, and metadata management in compliance with state and federal regulations.
- Ensure data security, privacy, and compliance standards (HIPAA, CJIS, FERPA, etc.) are met in all engineering processes.
- Support the integration of AI/ML pipelines, MLOps tools, and model versioning into production environments.
- Lead and mentor junior engineers, fostering a culture of collaboration and continuous improvement.
- Provide technical documentation and participate in technical reviews, architecture discussions, and state reporting requirements.
- Implement data interoperability and sharing pipelines that align with FAIR data principles (Findable, Accessible, Interoperable, and Reusable) to enhance transparency and collaboration across state agencies.
- Engineer and maintain metadata-driven architectures and APIs that enable discoverability, lineage tracking, and reproducibility of datasets in compliance with state governance requirements.
- Support development of synthetic data generation pipelines to enable privacy-preserving testing and innovation within sandbox environments prior to production deployment.
- Automate data validation, schema enforcement, and quality metrics reporting to ensure compliance with governance and audit requirements.
- Collaborate with governance, security, and AI ethics teams to ensure datasets adhere to state privacy laws, NIST SP 800-53, and ethical AI guidelines.
- Support phased rollout of data environments by developing repeatable, version-controlled deployment pipelines that allow for incremental testing and expansion.
Minimum Qualifications
- Bachelor’s or Master’s degree in Computer Science, Data Engineering, Information Systems, or related field.
- 8+ years of experience in data engineering, including design and deployment of production-grade data systems.
- Expertise in Python, SQL, and modern data engineering frameworks (e.g., PySpark, dbt, Airflow, Kafka).
- Hands-on experience with cloud platforms (AWS, Azure, or GCP) and data warehousing technologies (Snowflake, BigQuery, Redshift, Synapse).
- Strong understanding of data modeling, data governance, and API-based data integration.
- Experience supporting AI/ML initiatives, MLOps pipelines, and scalable data science environments.
- Demonstrated experience implementing FAIR-aligned data architectures and metadata cataloging within public-sector or research ecosystems.
- Proven ability to design and maintain compliant data-sharing pipelines that support sandbox testing and phased rollout strategies.
- Familiarity with accessibility and inclusivity requirements (WCAG 2.1, multilingual readiness) for data systems and visualization tools.
- Familiarity with public sector data standards, compliance frameworks, and procurement environments.
- US Citizenship or Permanent Residency required.
- Must reside in the Continental US.
- Depending on the government agency, specific requirements may include public trust background check or security clearance.
Preferred Qualifications
- Experience working on government contracts or with state data modernization programs.
- Familiarity with AI ethics, transparency, and bias mitigation practices.
- Experience with infrastructure as code (IaC) tools (Terraform, CloudFormation) and CI/CD pipelines.
- Knowledge of geospatial data systems, real-time data processing, or data API services.
- Experience in engineering or integrating synthetic data frameworks that support privacy-preserving research and model validation.
- Knowledge of metadata standards such as DCAT, schema.org, or CKAN to enable FAIR data publication and interoperability across systems.
- Familiarity with NIST AI RMF or ISO/IEC 5259 frameworks for AI data quality, lineage, and governance.
Top Skills
What We Do
We are your human-centric innovation partner, crafting impactful user experiences for government and commercial sectors with an unwavering commitment to efficiency and excellence.
Our Core Capabilities:
CMS Data Experience
Our team has established deep domain experience in insurance, claims, and quality data throughout the CMS ecosystem. We equip healthcare stakeholders with insights to make informed policy decisions through strategic data management, integrating diverse data sources, and conducting analysis.
Enabling Interoperability in Healthcare
We build functionality using standards such as FHIR to allow for data interoperability between government agencies, providers, and patients. Guided by our long-time partnership with CMS and healthcare stakeholders, we advocate for patient data accessibility, striving to enhance the quality of care delivered.
Leaders in DevSecOps
As some of the earliest practitioners of DevSecOps within CMS, we exemplify the significance of a carefully engineered CI/CD pipeline. Amidst unprecedented HealthCare.gov registrations, our team safeguards its stability and reliability.
Consulting to Innovate
At Element, we act as consultative partners at the crossroads of innovation and our clients' needs. No matter the challenge, we identify and prioritize the appropriate methods, ensuring tailored solutions and using modern technologies.
We believe in a world where solutions we build improve the lives of those who use them, and the elements of Innovation, Collaboration, Integrity, Equity, Creativity are the foundation








