Xealth
At Xealth, we’re revolutionizing healthcare by leveraging data and automation to empower care providers (building on EHRs such as Epic and Cerner) to seamlessly prescribe, deliver, and monitor digital health for patients. We are a detail-oriented team, committed to maintaining the highest standards while moving with agility and impact.
We are a highly skilled, collaborative, and passionate group, applying our expertise to improve health outcomes for millions. We believe in shared ownership and are looking for a team player who is a self-starter and self-driven to pioneer the next generation of intelligent, automated data insights.
This role offers a unique opportunity to join a data engineering team to advance our capabilities with data processing pipelines and our analytics product offering. There is a strong preference for this person to sit in the Seattle office; however, we are open to candidates in other locations within the United States.
What You’ll Own and Deliver (Responsibilities)
Data Modeling:
Execute expert-level Data Modeling and Design, utilizing dimensional modeling and denormalization techniques specifically for analytic workloads. Data Ingestion:
Ability to consume and process high-volume bounded and unbounded data, build robust Change Data Capture (CDC) mechanisms, and gather data from API calls and webhooks. Pipeline Design & Orchestration:
Design, build, and optimize high-volume, real-time streaming data pipelines using PySpark and Databricks environments. Scalability & Maintenance:
Maintain and scale large data lake pipelines, ensuring high performance and cost‑efficiency. Unit Testing & Quality Assurance:
Write comprehensive unit and integration tests for data pipelines to ensure code quality and production reliability. Cross-Functional Collaboration:
Partner with product managers and EHR specialists to translate clinical user behaviors into rich, analytical datasets, unlocking critical insights that drive evidence‑based improvements in healthcare processes. Technical Leadership:
Contribute to code reviews, system design discussions, and technical decisions that raise the engineering bar across the team. Automation and AI in Development:
Use AI‑assisted coding tools like GitHub Copilot to streamline development, increase quality, and accelerate delivery. The Expertise You’ll Bring (Requirements)
We’re looking for a data engineer with strong computer science fundamentals. Someone who’s comfortable reasoning about systems, data, and code structure at scale, and who’s excited to apply those skills in healthcare. CS Fundamentals:
Deep understanding of algorithms and data structures, with a specific focus on distributed computing principles (concurrency, partitioning, shuffling) necessary for processing large-scale datasets. Optimization & Troubleshooting:
Proficient in diagnosing complex failures in distributed processing jobs (e.g., Spark executor errors, memory leaks, data skew) using logs, distributed tracing, and performance metrics. Modern SQL and NoSQL Database Design:
Deep practical knowledge of open table formats, such as Delta Lake. Proficiency with common big data file formats, including Apache Parquet and Apache Avro. Infrastructure as Code (IaC):
Experience implementing IaC principles and tools for the automated deployment and management of data pipelines. API & Integration:
Hands‑on experience designing robust data ingestion frameworks via RESTful APIs and building event‑driven architectures for real‑time data flow. Cloud & Distributed Systems:
Experience designing and scaling cloud‑native data platforms and orchestrating data workloads using AWS and Kubernetes. Highly Valued Experience (Nice to Have)
Security & Governance:
Prior experience in a regulated industry with high security requirements. A good working understanding of data security principles, particularly regarding PHI and sensitive data governance. Real Time Stream Processing:
Expertise building streaming data pipelines, leveraging stream processors such as Apache Kafka and Apache Flink. Observability:
Experience implementing and utilizing data observability tools and practices to monitor data quality, lineage, and pipeline health. Visualization:
Experience building dashboards and visualizations to communicate data insights effectively. Benefits
Paid parental leave. Comprehensive medical, dental, and vision policies. Xealth covers 100% of employee premiums. We also provide Employee Assistance Programs. Xealth provides your laptop and offers a home office stipend. Generous learning & development opportunities for you to grow your skills and career. 401k match: Xealth offers a dollar‑for‑dollar match up to 3%. Flexible time off & 10 standardized holidays. $500 yearly fitness stipend to spend on staying active. Equal Employment Opportunity and Accommodations
Xealth is committed to providing reasonable accommodations for qualified individuals with disabilities and disabled veterans in our job application procedures. As set forth in Xealth’s Equal Employment Opportunity policy, we do not discriminate on the basis of any protected group status under any applicable law. Xealth’s mission is to improve patient care through data and automation.
#J-18808-Ljbffr
Data Modeling:
Execute expert-level Data Modeling and Design, utilizing dimensional modeling and denormalization techniques specifically for analytic workloads. Data Ingestion:
Ability to consume and process high-volume bounded and unbounded data, build robust Change Data Capture (CDC) mechanisms, and gather data from API calls and webhooks. Pipeline Design & Orchestration:
Design, build, and optimize high-volume, real-time streaming data pipelines using PySpark and Databricks environments. Scalability & Maintenance:
Maintain and scale large data lake pipelines, ensuring high performance and cost‑efficiency. Unit Testing & Quality Assurance:
Write comprehensive unit and integration tests for data pipelines to ensure code quality and production reliability. Cross-Functional Collaboration:
Partner with product managers and EHR specialists to translate clinical user behaviors into rich, analytical datasets, unlocking critical insights that drive evidence‑based improvements in healthcare processes. Technical Leadership:
Contribute to code reviews, system design discussions, and technical decisions that raise the engineering bar across the team. Automation and AI in Development:
Use AI‑assisted coding tools like GitHub Copilot to streamline development, increase quality, and accelerate delivery. The Expertise You’ll Bring (Requirements)
We’re looking for a data engineer with strong computer science fundamentals. Someone who’s comfortable reasoning about systems, data, and code structure at scale, and who’s excited to apply those skills in healthcare. CS Fundamentals:
Deep understanding of algorithms and data structures, with a specific focus on distributed computing principles (concurrency, partitioning, shuffling) necessary for processing large-scale datasets. Optimization & Troubleshooting:
Proficient in diagnosing complex failures in distributed processing jobs (e.g., Spark executor errors, memory leaks, data skew) using logs, distributed tracing, and performance metrics. Modern SQL and NoSQL Database Design:
Deep practical knowledge of open table formats, such as Delta Lake. Proficiency with common big data file formats, including Apache Parquet and Apache Avro. Infrastructure as Code (IaC):
Experience implementing IaC principles and tools for the automated deployment and management of data pipelines. API & Integration:
Hands‑on experience designing robust data ingestion frameworks via RESTful APIs and building event‑driven architectures for real‑time data flow. Cloud & Distributed Systems:
Experience designing and scaling cloud‑native data platforms and orchestrating data workloads using AWS and Kubernetes. Highly Valued Experience (Nice to Have)
Security & Governance:
Prior experience in a regulated industry with high security requirements. A good working understanding of data security principles, particularly regarding PHI and sensitive data governance. Real Time Stream Processing:
Expertise building streaming data pipelines, leveraging stream processors such as Apache Kafka and Apache Flink. Observability:
Experience implementing and utilizing data observability tools and practices to monitor data quality, lineage, and pipeline health. Visualization:
Experience building dashboards and visualizations to communicate data insights effectively. Benefits
Paid parental leave. Comprehensive medical, dental, and vision policies. Xealth covers 100% of employee premiums. We also provide Employee Assistance Programs. Xealth provides your laptop and offers a home office stipend. Generous learning & development opportunities for you to grow your skills and career. 401k match: Xealth offers a dollar‑for‑dollar match up to 3%. Flexible time off & 10 standardized holidays. $500 yearly fitness stipend to spend on staying active. Equal Employment Opportunity and Accommodations
Xealth is committed to providing reasonable accommodations for qualified individuals with disabilities and disabled veterans in our job application procedures. As set forth in Xealth’s Equal Employment Opportunity policy, we do not discriminate on the basis of any protected group status under any applicable law. Xealth’s mission is to improve patient care through data and automation.
#J-18808-Ljbffr