Logo
ADN Group

Solutions Architect – Data

ADN Group, Phoenix, Arizona, United States, 85003

Save Job

Job Title

Solutions Architect – Data

Location

Phoenix, AZ

Work Type

Onsite

Duration

06 Months of Contract

LinkedIn Profile Requirement

Required for Submission

Note

NO H1B // Local Candidates with DL

Primary Skills Solutions Architecture

Cloud Data Architecture

Azure

AWS

Databricks

ETL / ELT Pipeline Design

Data Lake Architecture

Python

SQL

Azure Data Factory (ADF)

Data Governance & Data Quality

Big Data Technologies (Spark, Hadoop, Kafka)

AI Architecture (RAG, Prompt Engineering)

CI/CD

DevOps

Git

Cloud Migration

Stakeholder & Cross-Team Collaboration

Job Overview

The Solutions Architect - Data is responsible for contributing to the design, modernization, optimization, and ongoing operations of enterprise-scale data systems for CHP. This role focuses on designing and implementing data solutions that organize, store, and manage data within a cloud-based data platform.

The architect will perform continuous maintenance and operational support within the cloud environment, including reviewing existing data infrastructure, planning future database solutions, and implementing systems that support data management needs for CHP users.

This role is also accountable for ensuring data integrity and governance, ensuring adherence to standards that maintain accuracy, consistency, and reliability across systems. The architect will identify data quality issues, analyze discrepancies, and drive resolution efforts.

The position requires a strong balance of architectural leadership, technical expertise, and collaboration with business stakeholders, data engineers, machine learning practitioners, and domain experts to deliver scalable, secure, and reliable AI-driven solutions.

The ideal candidate will have demonstrated experience delivering end-to-end ETL/ELT pipelines across Databricks, Azure, and AWS environments.

Key Responsibilities Design scalable data lake and enterprise data architectures using Databricks and cloud-native services

Develop metadata-driven, parameterized ingestion frameworks and multi-layer data architectures

Optimize data workloads and system performance

Define and enforce data governance frameworks for CHP

Design and develop reliable and scalable data pipelines

Architect AI systems, including RAG workflows and prompt engineering solutions

Lead cloud migration initiatives from legacy systems to modern data platforms

Provide architectural guidance, technical leadership, and best practices across teams

Create documentation, reusable components, and standardized architectural patterns

Required Skills and Experience Strong expertise with cloud platforms, primarily Azure or AWS

Hands‑on experience with Databricks

Strong proficiency in Python and SQL

Expertise in building ETL/ELT pipelines and ADF workflows

Experience designing data lakes and implementing data governance frameworks

Hands‑on experience with CI/CD, DevOps, and Git‑based development

Ability to translate business requirements into technical and architectural solutions

Technical Expertise Programming: Python, SQL, R

Big Data: Hadoop, Spark, Kafka, Hive

Cloud Platforms: Azure (ADF, Databricks, Azure OpenAI), AWS

Data Warehousing: Redshift, SQL Server

ETL/ELT Tools: SSIS

Education & Experience

Bachelor’s degree in Computer Science, Information Technology, Information Systems, Engineering, or a related field.

6+ years of experience in data engineering or .NET development.

Seniority Level

Mid‑Senior level

Employment Type

Contract

Job Function

Information Technology

Industries

IT Services and IT Consulting

#J-18808-Ljbffr