Logo
Dechen Consulting

Data Engineering Engineer

Dechen Consulting, Dearborn, Michigan, United States, 48120

Save Job

About Dechen Consulting Group (DCG)

Dechen Consulting Group (DCG) is a rapidly expanding, innovative IT Professional Services and Management Consulting company with a track record of more than twenty-five years in delivering skilled professionals to our clients across diverse sectors.

Job Opportunity

We are currently seeking a professional for a W2 contract opportunity in Dearborn, MI. This role has the potential to extend over multiple years, with the chance to transition to a direct hire position with our client. We provide healthcare, vacation, relocation assistance, and visa sponsorship/transfer. This is a W2 position, not C2C. THIRD PARTIES NEED NOT APPLY. This role offers excellent prospects for career progression!

Position Description

Employees in this job function are responsible for designing, building, and maintaining data solutions including data infrastructure, pipelines, etc. for collecting, storing, processing, and analyzing large volumes of data efficiently and accurately.

Key Responsibilities Collaborate with business and technology stakeholders to understand current and future data requirements. Design, build and maintain reliable, efficient, and scalable data infrastructure for data collection, storage, transformation, and analysis. Plan, design, build and maintain scalable data solutions including data pipelines, data models, and applications for efficient and reliable data workflow. Design, implement and maintain existing and future data platforms like data warehouses, data lakes, data lakehouse, etc. for structured and unstructured data. Design and develop analytical tools, algorithms, and programs to support data engineering activities like writing scripts and automating tasks. Ensure optimum performance and identify improvement opportunities. Skills Required

Google Cloud Platform ETL Apache Spark Data Architecture Python SQL KAFKA Skills Preferred

Java Powershell Data Acquisition Data Analysis Data Collection Data Conversion Data Integrity Data/Analytics dashboards Experience Required

4 years of Data Engineering work experience

Experience Preferred

Data Pipeline Architecture & Development: Design, build, and maintain highly scalable, fault-tolerant, and performant data pipelines to ingest and process data from 10 siloed sources, including both structured and unstructured formats. ML-Driven ETL Implementation: Operationalize ETL pipelines for intelligent data ingestion, automated cataloging, and sophisticated normalization of diverse datasets. Unified Data Model Creation: Architect and implement a unified data model capable of connecting all relevant data elements across various sources, optimized for efficient querying and insight generation by AI agents and chatbot interfaces. Big Data Processing: Utilize advanced distributed processing frameworks (Apache Beam, Apache Spark, Google Cloud Dataflow) to handle large-scale data transformations and data flow. Cloud-Native Data Infrastructure: Leverage GCP services to build and manage robust data storage, processing, and orchestration layers. Data Quality, Governance & Security: Implement rigorous data quality gates, validation rules, bad record handling, and comprehensive logging. Ensure strict adherence to data security policies, IAM role management, and GCP perimeter security. Automation & Orchestration: Develop shell scripts, Cloud Build YAMLs, and utilize Cloud Scheduler/PubSub for E2E automation of data pipelines and infrastructure provisioning. Collaboration with AI/ML Teams: Work closely with AI/ML engineers, data scientists, and product managers to understand data requirements, integrate data solutions with multi-agentic systems, and optimize data delivery for chatbot functionalities. Testing & CI/CD: Implement robust testing strategies, maintain high code quality through active participation in Git/GitHub, perform code reviews, and manage CI/CD pipelines via Cloud Build. Performance Tuning & Optimization: Continuously monitor, optimize, and troubleshoot data pipelines and BQ performance using techniques like table partitioning, clustering, and sharding.

Education Required

Bachelor's Degree

Education Preferred

Certification Program

Additional Information

Original Duration: 365 Days

4 days in the office.

We Are a People-Focused Company with a deep emphasis on family values and look forward to working with you.