Logo
Darwoft

[Hiring] Data Engineer @Darwoft

Darwoft, Germantown, Ohio, United States

Save Job

Oct 20, 2025 - Darwoft is hiring a remote Data Engineer.

Salary: contractor agreement with payment in usd. Location: LATAM.

Location:

LATAM (Remote supporting US-based teams) Job Type:

Contractor (Full-Time, Remote) Project:

Data Platform Modernization Healthcare Sector Time Zone:

Aligned with GMT-3 (Argentina) English Level:

B2/C1

Get to Know Us At Darwoft, we build software that drives real change. But were more than just tech were people first. With a remote-first culture and a highly collaborative team spread across LATAM, we partner with global companies to co-create reliable, scalable, and impactful digital products.

Were currently working with a leading US-based healthtech platform, in a major transformation of their data pipeline ecosystem migrating legacy SQL logic into modern, scalable cloud-based infrastructure using DBT, Spark, Argo, and AWS.

Were Looking For a Senior Data Engineer (DBT + Spark + Argo) In this role, you will be at the core of a strategic data transformation initiative: converting monolithic SQL Server logic into a modular, testable DBT architecture, while integrating Spark for performance and Argo for orchestration. You will work with cutting-edge lakehouse formats like Apache Hudi, Parquet, and Iceberg, and enable real-time analytics through ElasticSearch integration.

If you're passionate about modern data engineering and want to work in a data-driven, cloud-native, healthcare-focused environment, this is the role for you.

What Youll Be Doing

Translate legacy

T-SQL logic

into modular, scalable DBT models powered by

Spark SQL

Build reusable and performant

data transformation pipelines

Develop testing frameworks to ensure

data accuracy and integrity

in DBT workflows

Design and orchestrate workflows using

Argo Workflows

and CI/CD pipelines with

Argo CD

Manage mock data and

reference datasets

(e.g., ICD-10, CPT), ensuring version control and governance

Implement

efficient storage/query strategies

using Apache

Hudi ,

Parquet , and

Iceberg

Integrate ElasticSearch for analytics by building APIs and pipelines to support indexing and querying

Collaborate with DevOps teams to optimize

S3 usage , enforce

data security , and ensure compliance

Work in Agile squads and participate in planning, estimation, and sprint reviews

What You Bring

Strong experience with

DBT

for data modeling, testing, and deployment

Hands‑on proficiency in

Spark SQL , including performance tuning

Solid programming skills in

Python

for automation and data manipulation

Familiarity with

Jinja templating

for building reusable DBT components

Practical experience with

data lake formats : Apache Hudi, Parquet, Iceberg

Expertise in

Argo Workflows

and

CI/CD integration with Argo CD

Deep understanding of

AWS S3

data storage, performance tuning, and cost optimization

Strong command of

ElasticSearch

for indexing structured/unstructured data

Knowledge of

ICD-10 ,

CPT , and other healthcare data standards

Ability to work cross‑functionally in

Agile

environments

Nice to Have

Experience with

Docker ,

Kubernetes , and container orchestration

Familiarity with cloud‑native data tools:

AWS Glue ,

Databricks ,

EMR , or GCP equivalents

Prior work on CI/CD automation for data engineering workflows

Knowledge of

data compliance

standards: HIPAA, SOC2, etc.

Contributions to open‑source projects in DBT, Spark, or data engineering frameworks

Perks & Benefits

Contractor agreement with payment in USD

100% remote work

Argentinas public holidays

English classes

Referral program

Access to learning platforms

Explore this and other opportunities at: www.darwoft.com/careers

#J-18808-Ljbffr