Staff Data Engineer

Posted 12 Hours Ago
Easy Apply
Be an Early Applicant
Hiring Remotely in Venezuela
Remote
Senior level
Agency • Social Impact • Software • Consulting
Able is a product strategy and development company and we created a new model to help innovators realize their vision.
The Role
Design and build scalable data systems, manage data pipelines, ensure data quality, and support clinical data interoperability for a multi-country healthcare platform.
Summary Generated by Built In

Back in 2012, we were a group of engineers and designers who decided we wanted to build things—so we did. Able started as an engineering and product hub building for a portfolio of early-stage startups. We built many relationships while developing products that were thoughtful, effective, and genuinely useful. But, since then, we’ve grown… and so has our ambition. 

Now, we’re entering our next chapter—defined by applied AI. AI is a powerful force in the end-to-end software development cycle, and we’re creating practices that allow us to deliver software fast and more effectively than traditional approaches, creating meaningful value for our partners. Today, our builder mindset is driving us to become an AI-native organization across every function. We’re still evolving, and that’s part of the opportunity. If you want to build, learn, and tackle challenges alongside an ambitious team, let’s build together.

This position is 100% remote within LatAm.

About the Role

We’re looking for a Senior Data Engineer to design and build scalable data systems that power analytics and decision-making. You’ll define how data is captured, build reliable pipelines, and ensure data is accurate, accessible, and ready to use.

What We’re Looking ForDay-to-Day Responsibilities
  • Design, build, and operate a Databricks medallion lakehouse architecture (Bronze/Silver/Gold layers) using Delta Live Tables to support ingestion, transformation, and serving of clinical, behavioral, and operational data across a multi-country digital health platform
  • Architect and maintain scalable data pipelines on AWS (S3, Glue, Lambda, Kinesis, MSK/Kafka) that ingest data from diverse sources including FHIR-based clinical systems, remote patient monitoring devices, mobile applications, and third-party vendor APIs — ensuring reliability, idempotency, and observability at scale
  • Implement multi-country data isolation and governance leveraging Databricks Unity Catalog, enforcing data residency requirements across different countries (e.g., the US, EU, and the Kingdom of Saudi Arabia) and integrating policy-as-code consent enforcement (e.g., Open Policy Agent) aligned with regulatory requirements and guidelines (e.g., HIPAA, GDPR)
  • Partner with platform, compliance, and analytics teams to define and enforce data quality standards, lineage tracking, schema evolution strategies, and tamper-evident audit logging across all tiers of the lakehouse
  • Support clinical data interoperability by implementing and maintaining FHIR-to-OMOP mapping pipelines, enabling downstream analytics, population health reporting, and AI/ML feature engineering on harmonized datasets
  • Optimize data platform performance, cost, and reliability through partitioning strategies, compaction, caching, cluster sizing, and monitoring — targeting SLAs appropriate for a patient-facing healthcare platform operating at scale (e.g. 1M+ patients across a dozen markets)
  • Contribute to certification and compliance readiness (e.g., ISO 27001, SOC 2 Type 2) by maintaining documentation, change control processes, and validation artifacts for all data infrastructure components
  • Collaborate on real-time and event-driven architectures integrating Kafka-based streaming with the medallion layers and workflow orchestration, supporting adaptive patient journey logic and near-real-time analytics
Required Skills & Experience
  • Requires 8+ years of data engineering experience, with deep hands-on expertise in Databricks (Delta Lake, Unity Catalog, DLT), AWS data services, Python/Spark, and streaming frameworks — preferably within healthcare, life sciences, or other highly regulated industries
  • Strong proficiency with AWS data services such as S3, Glue, Lambda, Kinesis, Redshift, Athena, and IAM — with experience architecting end-to-end data pipelines in AWS-native or hybrid environments
  • Advanced Python and PySpark/Spark development skills for batch and streaming ETL/ELT pipeline development, data transformation, and data quality enforcement
  • Experience with streaming and event-driven architectures using Kafka (Amazon MSK or Confluent), including integration with lakehouse ingestion layers
  • Proven ability to implement data governance frameworks including data lineage, schema evolution, access controls, cataloging, and audit logging at enterprise scale
  • Strong understanding of data modeling for both analytical and operational use cases, including dimensional modeling, slowly changing dimensions, and schema-on-read patterns
  • Experience with infrastructure-as-code (Terraform, CloudFormation, or CDK) and CI/CD pipelines for data platform deployments
  • Familiarity with regulatory and compliance requirements in data management, including data residency, encryption at rest and in transit, and role-based access controls aligned with frameworks such as HIPAA, SOC 2, or ISO 27001
  • Excellent collaboration and communication skills, with the ability to work cross-functionally with platform engineering, analytics, clinical, and compliance teams
  • Bachelor's degree in Computer Science, Data Science, Engineering, or a related field (or equivalent practical experience)
Nice to Have
  • Experience with healthcare data standards and interoperability, particularly FHIR, OMOP CDM, HL7, and ICD-10, including building mapping or harmonization pipelines between clinical data models
  • Experience supporting certification and audit readiness (SOC 2 Type 2, ISO 27001) through documentation, change control, and validation of data infrastructure
  • Hands-on experience with multi-region or multi-tenant data architectures enforcing data sovereignty and residency requirements across multiple jurisdictions (e.g., US, EU, Middle East)
  • Familiarity with policy-as-code approaches to data access and consent enforcement, such as Open Policy Agent (OPA) or similar engines
  • Experience with workflow orchestration platforms and integrating data pipelines within broader application workflow contexts
  • Exposure to real-time analytics and ML feature engineering on top of lakehouse architectures, including feature stores and serving layers
  • Working knowledge of cost optimization strategies for Databricks and AWS, including cluster policies, spot instances, auto-scaling, and storage lifecycle management
  • Experience in life sciences, pharmaceutical, or regulated healthcare organizations, particularly in patient-facing digital health or chronic disease management contexts
  • Master's degree in a relevant technical field is a plus

Able is powered by curious, thoughtful people who care about what they build and how they build it. We’re actively investing in our team through AI training, knowledge-sharing, and hands-on experimentation to ensure everyone grows alongside the technology.

This position is 100% remote within LatAm. Strong verbal and written communication skills in English are a requirement. As a team member, you can expect:

  • To work 40 hours per week, and be available during normal business hours as needed
  • Payments made in USD
  • 18 days of PTO per year, observance of local holidays, and an annual break between Christmas and New Years
  • A monthly wellness stipend and snack boxes delivered to your home
About Able

Able builds technology products in a portfolio model. We believe that people, teams, and processes are more important than the ideas themselves, so we’ve focused on bringing great people together, and investing in their growth.

We’ve built products in a variety of industries. Everything from media to finance to toys to healthcare. Sometimes we work with management teams to help their businesses grow faster or unlock value using technology. Other times we start or buy businesses outright. Each time, we look for opportunities to leverage technology built at the portfolio-level to drive value faster.

Able is committed to inclusion and diversity and is an equal-opportunity employer. All applicants will receive consideration without regard to race, color, religion, gender, gender identity, sexual orientation, national origin, disability, or veteran status.

This is but the beginning of a conversation we’d love to have with you.

Apply, and let’s get this adventure started!

Top Skills

AWS
Cdk
CloudFormation
Databricks
Delta Lake
Kafka
Python
Spark
Terraform
Unity Catalog
Am I A Good Fit?
beta
Get Personalized Job Insights.
Our AI-powered fit analysis compares your resume with a job listing so you know if your skills & experience align.

The Company
HQ: New York, NY
87 Employees
Year Founded: 2013

What We Do

Able is a product development studio dedicated to accelerating positive outcomes for early-stage companies. With a decade of experience in healthtech, fintech, and social good, we’ve launched numerous successful ventures. Our team of 100 experts in engineering, design, and entrepreneurship is committed to shaping the future through innovative product development. We offer comprehensive services covering the entire software product development lifecycle—from market analysis to build execution. Whether it’s UX design, AI strategies, or groundbreaking software, we’re here to deliver impactful results. Plus, we’re friendly, collaborative, and ambitious!

Why Work With Us

Founded in 2013 by Mike Potter, a software engineer and entrepreneur, and Michael Polansky, a venture investor, Able was built on the premise that entrepreneurs and philanthropists need a better services model for building technology. Since then, we have partnered closely with select founders, creating bespoke teams with hard-to-find product, desig

Gallery

Gallery

Similar Jobs

Luxury Presence Logo Luxury Presence

Staff Data Engineer

Marketing Tech • Real Estate • Software • PropTech • SEO
Easy Apply
Remote or Hybrid
12 Locations
500 Employees
Remote
Venezuela
36 Employees

Luxury Presence Logo Luxury Presence

Senior Data Engineer

Marketing Tech • Real Estate • Software • PropTech • SEO
Easy Apply
Remote or Hybrid
12 Locations
500 Employees

Rubrik Logo Rubrik

Join Our Engineering Talent Community

Artificial Intelligence • Big Data • Cloud • Information Technology • Software • Cybersecurity • Data Privacy
Remote
Venezuela
3000 Employees

Similar Companies Hiring

Milestone Systems Thumbnail
Software • Security • Other • Big Data Analytics • Artificial Intelligence • Analytics
Lake Oswego, OR
1500 Employees
Fairly Even Thumbnail
Hardware • Other • Robotics • Sales • Software • Hospitality
New York, NY
30 Employees
Kepler  Thumbnail
Fintech • Software
New York, New York
6 Employees

Sign up now Access later

Create Free Account

Please log in or sign up to report this job.

Create Free Account