Staff Data Engineer @Able
Software Development
Salary unspecified
Remote Location
Job Type full-time
Posted YDay

[Hiring] Staff Data Engineer @Able

YDay - Able is hiring a remote Staff Data Engineer. πŸ’Έ Salary: unspecified πŸ“Location: Latin America (LATAM)

Role Description

We’re looking for a Senior Data Engineer to design and build scalable data systems that power analytics and decision-making. You’ll define how data is captured, build reliable pipelines, and ensure data is accurate, accessible, and ready to use.

Qualifications

  • Requires 8+ years of data engineering experience, with deep hands-on expertise in Databricks (Delta Lake, Unity Catalog, DLT), AWS data services, Python/Spark, and streaming frameworks β€” preferably within healthcare, life sciences, or other highly regulated industries
  • Strong proficiency with AWS data services such as S3, Glue, Lambda, Kinesis, Redshift, Athena, and IAM β€” with experience architecting end-to-end data pipelines in AWS-native or hybrid environments
  • Advanced Python and PySpark/Spark development skills for batch and streaming ETL/ELT pipeline development, data transformation, and data quality enforcement
  • Experience with streaming and event-driven architectures using Kafka (Amazon MSK or Confluent), including integration with lakehouse ingestion layers
  • Proven ability to implement data governance frameworks including data lineage, schema evolution, access controls, cataloging, and audit logging at enterprise scale
  • Strong understanding of data modeling for both analytical and operational use cases, including dimensional modeling, slowly changing dimensions, and schema-on-read patterns
  • Experience with infrastructure-as-code (Terraform, CloudFormation, or CDK) and CI/CD pipelines for data platform deployments
  • Familiarity with regulatory and compliance requirements in data management, including data residency, encryption at rest and in transit, and role-based access controls aligned with frameworks such as HIPAA, SOC 2, or ISO 27001
  • Excellent collaboration and communication skills, with the ability to work cross-functionally with platform engineering, analytics, clinical, and compliance teams
  • Bachelor's degree in Computer Science, Data Science, Engineering, or a related field (or equivalent practical experience)

Requirements

  • Design, build, and operate a Databricks medallion lakehouse architecture (Bronze/Silver/Gold layers) using Delta Live Tables to support ingestion, transformation, and serving of clinical, behavioral, and operational data across a multi-country digital health platform
  • Architect and maintain scalable data pipelines on AWS (S3, Glue, Lambda, Kinesis, MSK/Kafka) that ingest data from diverse sources including FHIR-based clinical systems, remote patient monitoring devices, mobile applications, and third-party vendor APIs β€” ensuring reliability, idempotency, and observability at scale
  • Implement multi-country data isolation and governance leveraging Databricks Unity Catalog, enforcing data residency requirements across different countries (e.g., the US, EU, and the Kingdom of Saudi Arabia) and integrating policy-as-code consent enforcement (e.g., Open Policy Agent) aligned with regulatory requirements and guidelines (e.g., HIPAA, GDPR)
  • Partner with platform, compliance, and analytics teams to define and enforce data quality standards, lineage tracking, schema evolution strategies, and tamper-evident audit logging across all tiers of the lakehouse
  • Support clinical data interoperability by implementing and maintaining FHIR-to-OMOP mapping pipelines, enabling downstream analytics, population health reporting, and AI/ML feature engineering on harmonized datasets
  • Optimize data platform performance, cost, and reliability through partitioning strategies, compaction, caching, cluster sizing, and monitoring β€” targeting SLAs appropriate for a patient-facing healthcare platform operating at scale (e.g. 1M+ patients across a dozen markets)
  • Contribute to certification and compliance readiness (e.g., ISO 27001, SOC 2 Type 2) by maintaining documentation, change control processes, and validation artifacts for all data infrastructure components
  • Collaborate on real-time and event-driven architectures integrating Kafka-based streaming with the medallion layers and workflow orchestration, supporting adaptive patient journey logic and near-real-time analytics

Benefits

  • To work 40 hours per week, and be available during normal business hours as needed
  • Payments made in USD
  • 18 days of PTO per year, observance of local holidays, and an annual break between Christmas and New Years
  • A monthly wellness stipend and snack boxes delivered to your home
Before You Apply
️
remote Be aware of the location restriction for this remote position: Latin America (LATAM)
β€Ό Beware of scams! When applying for jobs, you should NEVER have to pay anything. Learn more.
Staff Data Engineer @Able
Software Development
Salary unspecified
Remote Location
Job Type full-time
Posted YDay
Apply for this position
Did not apply βœ“
Applied βœ“
Sent Follow-Up βœ“
Interview Scheduled βœ“
Interview Completed βœ“
Offer Accepted βœ“
Offer Declined βœ“
Unlock 152,720 Remote Jobs
️
remote Be aware of the location restriction for this remote position: Latin America (LATAM)
β€Ό Beware of scams! When applying for jobs, you should NEVER have to pay anything. Learn more.
Apply for this position
Did not apply βœ“
Applied βœ“
Sent Follow-Up βœ“
Interview Scheduled βœ“
Interview Completed βœ“
Offer Accepted βœ“
Offer Declined βœ“
Unlock 152,720 Remote Jobs
Γ—

Apply to the best remote jobs
before everyone else

Access 152,720+ vetted remote jobs and get daily alerts.

4.9 β˜…β˜…β˜…β˜…β˜… from 500+ reviews
Unlock All Jobs Now

Maybe later