Logo
Acunor Infotech

Databricks Engineer – Commercial Insurance – Fulltime

Acunor Infotech, Columbus, Ohio, United States, 43224

Save Job

ob Title:

Data Engineer (Databricks Azure) Client:

One of our Consulting Clients (Global Analytics & Digital Transformation Firm) Location:

Columbus, OH (Remote/Hybrid) Duration:

Full-Time About the Role

We are seeking a highly skilled

Data Engineer

with deep expertise in

Databricks

and

Azure Cloud

to join a decision analytics and data engineering team within one of our global consulting clients. The role involves

building, optimizing, and maintaining large-scale data pipelines

that fuel enterprise analytics, reporting, and AI-driven insights-primarily supporting clients in the

insurance and financial services domains . Key Responsibilities

Data Pipeline Development & Optimization

Design, build, and enhance

ETL/ELT data pipelines

using

Azure Data Factory

,

Databricks (PySpark, SQL, Python) , and related services. Develop and manage

Delta Live Tables

,

Autoloader

, and

Unity Catalog

within the Databricks ecosystem for structured, incremental data processing. Implement

data ingestion, transformation, and validation frameworks

that ensure high performance, scalability, and reliability. Monitor data pipelines, troubleshoot issues, and ensure optimal system performance and SLA adherence. Data Modeling & Architecture

Collaborate with business analysts and reporting teams to define

logical and physical data models

supporting analytical and operational needs. Implement

data warehousing and lakehouse solutions

using

Azure Data Lake

and

Delta Lake . Optimize data structures for query performance, cost efficiency, and reusability. Data Quality, Governance & Automation

Design and implement robust

data quality checks

and validation mechanisms to maintain integrity across sources and transformations. Automate repetitive processes using scripts, parameterized pipelines, and reusable frameworks. Conduct periodic audits and compliance checks aligned with governance policies. Contribute to metadata management, documentation, and lineage tracking. Required Skills & Experience

7 12 years

of experience in

Data Engineering

with proven expertise in

Databricks

and

Azure Cloud

ecosystems. Strong hands‑on experience in

PySpark

,

Python

, and

SQL

for data transformation, validation, and performance tuning. Solid understanding of

Delta Lake architecture

,

ETL/ELT frameworks

, and

data warehousing

principles. Proficiency with

Azure services

including

Data Factory (ADF)

,

Data Lake (ADLS)

, and

Databricks Notebooks . Experience with

Delta Live Tables

,

Unity Catalog

, and

Autoloader

for batch and streaming data processing. Strong background in

data modeling

,

performance optimization

, and

automation scripting . Familiarity with

Agile methodologies

and DevOps-based deployment practices (Git, CI/CD preferred). Strong analytical, communication, and problem‑solving skills to collaborate effectively across diverse teams. Preferred: Exposure to

insurance, healthcare, or financial services

data ecosystems. Nice to Have

Experience in

data migration projects

(on‑prem to cloud or multi‑cloud). Familiarity with

Delta Sharing

,

Databricks SQL Warehouses

, or

MLflow

for advanced use cases. Experience with

data cataloging, lineage, or quality frameworks

such as

Purview ,

Collibra , or

Great Expectations . Exposure to

BI/reporting tools

like Power BI or Tableau for end‑to‑end integration understanding.

#J-18808-Ljbffr