P.A. Consulting
Job Description
This is a CONTRACT TO POSSIBLE HIRE POSITION!
Were seeking a
Mid-Level Data Engineer
with hands-on experience in configuring and optimizing
Apache Iceberg
infrastructure. The role involves building out our foundational Iceberg data lakehouse architecture and integrating it with key cloud and analytics platforms. You will be a core part of our data engineering team, working closely with analytics and BI teams to ensure seamless data access and usability. Key Responsibilities Iceberg Infrastructure Configuration Design and implement the initial Apache Iceberg infrastructure. Ensure compatibility and optimization for batch and streaming data use cases. Platform Integration & Connectivity Set up and manage connections between Iceberg and:
Google Cloud Platform (GCP) Snowflake
With a focus on
Federated Data Warehousing (FDW) . MicroStrategy Looker Power BI
(lower priority, but still considered for downstream enablement)
Data Pipeline Development Build and deploy initial pipelines for data flow from:
Snowflake ? Iceberg ? MicroStrategy
Monitor and optimize data ingestion, transformation, and delivery. Ensure data quality, lineage, and security compliance throughout the pipeline. Collaboration & Documentation Collaborate cross-functionally with data science, analytics, and DevOps teams. Document configuration, design patterns, and integration processes. Qualifications: Qualifications Required: 35 years of experience in data engineering or related field. Proven experience configuring and managing
Apache Iceberg
environments. Hands-on experience with
Snowflake
, including familiarity with
FDW . Experience integrating cloud storage systems and query engines (e.g., BigQuery, GCP). Working knowledge of BI tools:
MicroStrategy
,
Looker
,
Power BI . Proficiency in Python, SQL, and data orchestration tools (e.g., Airflow). Strong understanding of data lakehouse architecture and performance optimization. Preferred: Familiarity with secure data sharing and access control across tools. Knowledge of metadata catalogs such as Apache Hive, AWS Glue, or Unity Catalog. Background in working with distributed data systems and cloud-native environments.
#J-18808-Ljbffr
Mid-Level Data Engineer
with hands-on experience in configuring and optimizing
Apache Iceberg
infrastructure. The role involves building out our foundational Iceberg data lakehouse architecture and integrating it with key cloud and analytics platforms. You will be a core part of our data engineering team, working closely with analytics and BI teams to ensure seamless data access and usability. Key Responsibilities Iceberg Infrastructure Configuration Design and implement the initial Apache Iceberg infrastructure. Ensure compatibility and optimization for batch and streaming data use cases. Platform Integration & Connectivity Set up and manage connections between Iceberg and:
Google Cloud Platform (GCP) Snowflake
With a focus on
Federated Data Warehousing (FDW) . MicroStrategy Looker Power BI
(lower priority, but still considered for downstream enablement)
Data Pipeline Development Build and deploy initial pipelines for data flow from:
Snowflake ? Iceberg ? MicroStrategy
Monitor and optimize data ingestion, transformation, and delivery. Ensure data quality, lineage, and security compliance throughout the pipeline. Collaboration & Documentation Collaborate cross-functionally with data science, analytics, and DevOps teams. Document configuration, design patterns, and integration processes. Qualifications: Qualifications Required: 35 years of experience in data engineering or related field. Proven experience configuring and managing
Apache Iceberg
environments. Hands-on experience with
Snowflake
, including familiarity with
FDW . Experience integrating cloud storage systems and query engines (e.g., BigQuery, GCP). Working knowledge of BI tools:
MicroStrategy
,
Looker
,
Power BI . Proficiency in Python, SQL, and data orchestration tools (e.g., Airflow). Strong understanding of data lakehouse architecture and performance optimization. Preferred: Familiarity with secure data sharing and access control across tools. Knowledge of metadata catalogs such as Apache Hive, AWS Glue, or Unity Catalog. Background in working with distributed data systems and cloud-native environments.
#J-18808-Ljbffr