Diverse Lynx
Role Overview
Responsible for designing, developing, and maintaining
data pipelines into a Lakehouse architecture
to support analytics, reporting, and regulatory needs. Works extensively with
PySpark on Databricks , orchestrates workflows with
Apache Airflow , and leverages
AWS cloud services
for storage, compute, and security.
Key Responsibilities Design, Build and optimize
ETL/ELT pipelines
using PySpark on Databricks. Manage
Apache Airflow DAGs
for scheduling and workflow orchestration. Ingest and transform data into
Delta Lake (bronze/silver/gold layers) . Leverage
AWS services
(S3, EC2, Lambda, IAM) for data integration. Implement
data modeling, schema enforcement, and governance . Monitor and improve
pipeline reliability, performance, and cost efficiency . Required Skills
Proficiency in
PySpark and Databricks
(Delta Lake, clusters, jobs). Hands-on with
Apache Airflow
(DAG design, monitoring). Strong in
AWS services : S3, EC2, Lambda, IAM. Strong SQL and
Python
for transformations and orchestration. Knowledge of
Lakehouse architecture (Delta Lake)
and data modeling. Experience in
ETL/ELT and data warehousing best practices .
Diverse Lynx LLC is an Equal Employment Opportunity employer. All qualified applicants will receive due consideration for employment without any discrimination. All applicants will be evaluated solely on the basis of their ability, competence and their proven capability to perform the functions outlined in the corresponding role. We promote and support a diverse workforce across all levels in the company.
Responsible for designing, developing, and maintaining
data pipelines into a Lakehouse architecture
to support analytics, reporting, and regulatory needs. Works extensively with
PySpark on Databricks , orchestrates workflows with
Apache Airflow , and leverages
AWS cloud services
for storage, compute, and security.
Key Responsibilities Design, Build and optimize
ETL/ELT pipelines
using PySpark on Databricks. Manage
Apache Airflow DAGs
for scheduling and workflow orchestration. Ingest and transform data into
Delta Lake (bronze/silver/gold layers) . Leverage
AWS services
(S3, EC2, Lambda, IAM) for data integration. Implement
data modeling, schema enforcement, and governance . Monitor and improve
pipeline reliability, performance, and cost efficiency . Required Skills
Proficiency in
PySpark and Databricks
(Delta Lake, clusters, jobs). Hands-on with
Apache Airflow
(DAG design, monitoring). Strong in
AWS services : S3, EC2, Lambda, IAM. Strong SQL and
Python
for transformations and orchestration. Knowledge of
Lakehouse architecture (Delta Lake)
and data modeling. Experience in
ETL/ELT and data warehousing best practices .
Diverse Lynx LLC is an Equal Employment Opportunity employer. All qualified applicants will receive due consideration for employment without any discrimination. All applicants will be evaluated solely on the basis of their ability, competence and their proven capability to perform the functions outlined in the corresponding role. We promote and support a diverse workforce across all levels in the company.