Darwoft
Oct 20, 2025 - Darwoft is hiring a remote Data Engineer.
Salary: contractor agreement with payment in usd. Location: LATAM.
Location:
LATAM (Remote supporting US-based teams) Job Type:
Contractor (Full-Time, Remote) Project:
Data Platform Modernization Healthcare Sector Time Zone:
Aligned with GMT-3 (Argentina) English Level:
B2/C1
Get to Know Us At Darwoft, we build software that drives real change. But were more than just tech were people first. With a remote-first culture and a highly collaborative team spread across LATAM, we partner with global companies to co-create reliable, scalable, and impactful digital products.
Were currently working with a leading US-based healthtech platform, in a major transformation of their data pipeline ecosystem migrating legacy SQL logic into modern, scalable cloud-based infrastructure using DBT, Spark, Argo, and AWS.
Were Looking For a Senior Data Engineer (DBT + Spark + Argo) In this role, you will be at the core of a strategic data transformation initiative: converting monolithic SQL Server logic into a modular, testable DBT architecture, while integrating Spark for performance and Argo for orchestration. You will work with cutting-edge lakehouse formats like Apache Hudi, Parquet, and Iceberg, and enable real-time analytics through ElasticSearch integration.
If you're passionate about modern data engineering and want to work in a data-driven, cloud-native, healthcare-focused environment, this is the role for you.
What Youll Be Doing
Translate legacy
T-SQL logic
into modular, scalable DBT models powered by
Spark SQL
Build reusable and performant
data transformation pipelines
Develop testing frameworks to ensure
data accuracy and integrity
in DBT workflows
Design and orchestrate workflows using
Argo Workflows
and CI/CD pipelines with
Argo CD
Manage mock data and
reference datasets
(e.g., ICD-10, CPT), ensuring version control and governance
Implement
efficient storage/query strategies
using Apache
Hudi ,
Parquet , and
Iceberg
Integrate ElasticSearch for analytics by building APIs and pipelines to support indexing and querying
Collaborate with DevOps teams to optimize
S3 usage , enforce
data security , and ensure compliance
Work in Agile squads and participate in planning, estimation, and sprint reviews
What You Bring
Strong experience with
DBT
for data modeling, testing, and deployment
Hands‑on proficiency in
Spark SQL , including performance tuning
Solid programming skills in
Python
for automation and data manipulation
Familiarity with
Jinja templating
for building reusable DBT components
Practical experience with
data lake formats : Apache Hudi, Parquet, Iceberg
Expertise in
Argo Workflows
and
CI/CD integration with Argo CD
Deep understanding of
AWS S3
data storage, performance tuning, and cost optimization
Strong command of
ElasticSearch
for indexing structured/unstructured data
Knowledge of
ICD-10 ,
CPT , and other healthcare data standards
Ability to work cross‑functionally in
Agile
environments
Nice to Have
Experience with
Docker ,
Kubernetes , and container orchestration
Familiarity with cloud‑native data tools:
AWS Glue ,
Databricks ,
EMR , or GCP equivalents
Prior work on CI/CD automation for data engineering workflows
Knowledge of
data compliance
standards: HIPAA, SOC2, etc.
Contributions to open‑source projects in DBT, Spark, or data engineering frameworks
Perks & Benefits
Contractor agreement with payment in USD
100% remote work
Argentinas public holidays
English classes
Referral program
Access to learning platforms
Explore this and other opportunities at: www.darwoft.com/careers
#J-18808-Ljbffr
Salary: contractor agreement with payment in usd. Location: LATAM.
Location:
LATAM (Remote supporting US-based teams) Job Type:
Contractor (Full-Time, Remote) Project:
Data Platform Modernization Healthcare Sector Time Zone:
Aligned with GMT-3 (Argentina) English Level:
B2/C1
Get to Know Us At Darwoft, we build software that drives real change. But were more than just tech were people first. With a remote-first culture and a highly collaborative team spread across LATAM, we partner with global companies to co-create reliable, scalable, and impactful digital products.
Were currently working with a leading US-based healthtech platform, in a major transformation of their data pipeline ecosystem migrating legacy SQL logic into modern, scalable cloud-based infrastructure using DBT, Spark, Argo, and AWS.
Were Looking For a Senior Data Engineer (DBT + Spark + Argo) In this role, you will be at the core of a strategic data transformation initiative: converting monolithic SQL Server logic into a modular, testable DBT architecture, while integrating Spark for performance and Argo for orchestration. You will work with cutting-edge lakehouse formats like Apache Hudi, Parquet, and Iceberg, and enable real-time analytics through ElasticSearch integration.
If you're passionate about modern data engineering and want to work in a data-driven, cloud-native, healthcare-focused environment, this is the role for you.
What Youll Be Doing
Translate legacy
T-SQL logic
into modular, scalable DBT models powered by
Spark SQL
Build reusable and performant
data transformation pipelines
Develop testing frameworks to ensure
data accuracy and integrity
in DBT workflows
Design and orchestrate workflows using
Argo Workflows
and CI/CD pipelines with
Argo CD
Manage mock data and
reference datasets
(e.g., ICD-10, CPT), ensuring version control and governance
Implement
efficient storage/query strategies
using Apache
Hudi ,
Parquet , and
Iceberg
Integrate ElasticSearch for analytics by building APIs and pipelines to support indexing and querying
Collaborate with DevOps teams to optimize
S3 usage , enforce
data security , and ensure compliance
Work in Agile squads and participate in planning, estimation, and sprint reviews
What You Bring
Strong experience with
DBT
for data modeling, testing, and deployment
Hands‑on proficiency in
Spark SQL , including performance tuning
Solid programming skills in
Python
for automation and data manipulation
Familiarity with
Jinja templating
for building reusable DBT components
Practical experience with
data lake formats : Apache Hudi, Parquet, Iceberg
Expertise in
Argo Workflows
and
CI/CD integration with Argo CD
Deep understanding of
AWS S3
data storage, performance tuning, and cost optimization
Strong command of
ElasticSearch
for indexing structured/unstructured data
Knowledge of
ICD-10 ,
CPT , and other healthcare data standards
Ability to work cross‑functionally in
Agile
environments
Nice to Have
Experience with
Docker ,
Kubernetes , and container orchestration
Familiarity with cloud‑native data tools:
AWS Glue ,
Databricks ,
EMR , or GCP equivalents
Prior work on CI/CD automation for data engineering workflows
Knowledge of
data compliance
standards: HIPAA, SOC2, etc.
Contributions to open‑source projects in DBT, Spark, or data engineering frameworks
Perks & Benefits
Contractor agreement with payment in USD
100% remote work
Argentinas public holidays
English classes
Referral program
Access to learning platforms
Explore this and other opportunities at: www.darwoft.com/careers
#J-18808-Ljbffr