Logo
Diverse Lynx

Data Engineer

Diverse Lynx, Dallas, North Carolina, United States, 28034

Save Job

Role Overview

Responsible for designing, developing, and maintaining

data pipelines into a Lakehouse architecture

to support analytics, reporting, and regulatory needs. Works extensively with

PySpark on Databricks , orchestrates workflows with

Apache Airflow , and leverages

AWS cloud services

for storage, compute, and security.

Key Responsibilities Design, Build and optimize

ETL/ELT pipelines

using PySpark on Databricks. Manage

Apache Airflow DAGs

for scheduling and workflow orchestration. Ingest and transform data into

Delta Lake (bronze/silver/gold layers) . Leverage

AWS services

(S3, EC2, Lambda, IAM) for data integration. Implement

data modeling, schema enforcement, and governance . Monitor and improve

pipeline reliability, performance, and cost efficiency . Required Skills

Proficiency in

PySpark and Databricks

(Delta Lake, clusters, jobs). Hands-on with

Apache Airflow

(DAG design, monitoring). Strong in

AWS services : S3, EC2, Lambda, IAM. Strong SQL and

Python

for transformations and orchestration. Knowledge of

Lakehouse architecture (Delta Lake)

and data modeling. Experience in

ETL/ELT and data warehousing best practices .

Diverse Lynx LLC is an Equal Employment Opportunity employer. All qualified applicants will receive due consideration for employment without any discrimination. All applicants will be evaluated solely on the basis of their ability, competence and their proven capability to perform the functions outlined in the corresponding role. We promote and support a diverse workforce across all levels in the company.