Logo
RELI Group, Inc.

Databricks Engineer

RELI Group, Inc., Woodlawn, Maryland, United States

Save Job

Job Location:

Windsor Mill, MD Remote Type:

Fully Remote Position Type:

Full Time Education Level:

4 Year Degree Salary Range:

$100000.00 - $150000.00

About Us RELI Group, our work is grounded in purpose. We partner with government agencies to solve complex challenges, improve public health, strengthen national security, and make government services more effective and efficient. Our team brings deep expertise and a shared commitment to delivering meaningful outcomes. Behind every solution is a group of experts who care deeply about impact.

Position Summary RELI Group is seeking a highly skilled Data Engineer to support the Centers for Medicare & Medicaid Services (CMS) Multidimensional Information Data Analytics System (MIDAS) Program. This role will play a critical part in the modernization of the MIDAS data platform, building and maintaining scalable, high-performing data pipelines that drive enterprise analytics, reporting, and operational decision-making. The Data Engineer will focus on developing data ingestion, transformation, and processing pipelines across AWS and Databricks-based environments, with an emphasis on Delta Lake lakehouse architectures.

Responsibilities

Design, build, and maintain robust ETL/ELT pipelines that ingest, transform, and curate data for analytics and reporting solutions.

Develop efficient data processing workflows using Python, SQL, and PySpark within Databricks and Delta Lake environments.

Implement scalable data lakehouse architectures supporting complex healthcare datasets and enterprise analytics.

Collaborate with Data Architects, Automation Test Engineers, QA Analysts, and Business Analysts to define data transformation logic and data integration standards.

Build reusable code and frameworks for data ingestion, data quality validation, and exception handling.

Optimize data processing jobs for performance, reliability, scalability, and cost efficiency in AWS and Databricks cloud platforms.

Participate in schema design, data modeling, and version-controlled data pipeline development.

Contribute to data governance, metadata management, and data lineage documentation to support audit, compliance, and FISMA reporting requirements.

Partner with DevOps teams to integrate data pipelines into CI/CD workflows using GitHub, Databricks Repos, and related tools.

Perform root cause analysis and resolution of data issues across staging, integration, and production environments.

Qualifications

Bachelor's degree in Computer Science, Information Systems, Engineering, or related technical field

6+ years of experience in data engineering or data pipeline development roles

Strong hands-on experience developing data pipelines using Python, SQL, and PySpark

Experience in Java is a plus, to understand existing code and convert to Python and Databricks Notebooks

Extensive experience with Databricks, including Databricks Workflows, Delta Lake, notebooks, and distributed computing

Proven experience with ETL/ELT design, development, and optimization for large-scale data processing

Solid understanding of data lakehouse architecture, data partitioning, and Delta Lake versioning

Experience with AWS data services such as S3, Redshift, Glue, IAM, and RDS

Familiarity with CI/CD processes for data pipelines (GitHub Actions, Jenkins, Databricks Repos)

Experience with data quality validation, data profiling, and debugging data pipeline failures

Strong collaboration and communication skills with ability to translate business needs into technical requirements

Preferred Qualifications

Experience supporting CMS programs and understanding of the CMS Technical Reference Architecture (TRA) and Target Lifecycle (TLC).

Exposure to BI platforms such as QuickSight, Tableau, or Power BI for data consumption and reporting validation.

Familiarity with Unity Catalog, data governance, and access control within Databricks.

Experience supporting regulated environments requiring audit trails, FISMA audits, or CMS Acceptable Risk Safeguards (ARS 5.0).

Exposure to healthcare datasets such as ACA, QHP, HICS, or HIOS.

Summary of Core Technologies

Python, SQL, PySpark

Databricks, Delta Lake, Databricks Workflows

AWS: S3, Redshift, Glue, RDS

CI/CD: GitHub Actions, Jenkins, Databricks Repos

ETL/ELT pipelines and data lakehouse transformations

Distributed data processing and optimization

EEO Employer RELI Group is an Equal Employment Opportunity / Affirmative Action employer. All qualified applicants will receive consideration for employment without regard to race, color, national origin, ancestry, citizenship status, military status, protected veteran status, religion, creed, physical or mental disability, medical condition, marital status, sex, sexual orientation, gender, gender identity or expression, age, genetic information, or any other basis protected by law, ordinance, or regulation.

HUBZone We encourage all candidates who live in a HUBZone to apply. You can check to see if your address is located in a HUBZone by accessing the SBA HUBZone Map.

The annual salary range for this position is $100,000 to $150,000. Actual compensation will depend on a range of factors, including but not limited to the individual's skills, experience, qualifications, certifications, location, and applicable employment laws. RELI Group provides a variety of additional benefits to its employees.

#J-18808-Ljbffr