Logo
Tandym Tech

Data Engineer

Tandym Tech, Irvine, California, United States, 92713

Save Job

A recognized property management company in California is actively seeking a new Data Engineer to join their growing team. In this role, the Data Engineer will support the design, automation, and operation of modern data infrastructure.

Responsibilities

Design, build, and maintain scalable and resilient CI/CD pipelines for data applications and infrastructure, with a strong focus on Snowflake, dbt, and modern data tooling

Implement and manage Snowflake dbt projects, including developing dbt models, tests, and documentation, and integrating dbt workflows into CI/CD pipelines

Develop and manage Infrastructure as Code (IaC) using Terraform to provision and configure data infrastructure on Google Cloud Platform (Google Cloud Platform)

Automate the deployment, monitoring, and management of Snowflake data warehouse environments, ensuring optimal performance, security, reliability, and cost efficiency

Collaborate with data engineers and data scientists to understand requirements and deliver automated solutions for data ingestion, transformation, and delivery

Implement and maintain monitoring, logging, and alerting for data pipelines and infrastructure to ensure high availability and proactive issue resolution

Develop and maintain automation scripts and tooling using Python (primary) and Bash for operational tasks

Apply and maintain security best practices across data infrastructure, pipelines, and CI/CD processes

Troubleshoot and resolve issues related to data pipelines, infrastructure, and deployments

Participate in code reviews for Terraform, dbt models, and automation scripts

Create and maintain clear technical documentation for architectures, configurations, and operational processes

Perform other duties, as needed

Qualifications

5+ years of experience in Data Engineering, Analytics Engineering, or DevOps for data platforms

Strong hands-on experience with Snowflake in production environments

Proven expertise with dbt, including model development, testing, and documentation

Experience building and maintaining CI/CD pipelines for data and cloud infrastructure

Hands-on experience with Terraform and Infrastructure as Code practices

Experience working with Google Cloud Platform (Google Cloud Platform) for data storage, processing, and analytics

Strong programming skills in Python and working knowledge of Bash/shell scripting

Experience implementing monitoring, logging, and alerting for data systems

Solid understanding of data pipeline architectures and modern data stack concepts

Strong troubleshooting and problem-solving skills

Ability to collaborate effectively with cross-functional technical teams

Desired Skills

Experience with Data Governance and Data Quality frameworks

Familiarity with Cost Optimization strategies in Snowflake and cloud environments

Experience supporting High-Availability or Mission-Critical Data platforms

Prior contract or consulting experience in fast-paced environments

#J-18808-Ljbffr