Tandym Tech
A recognized property management company in California is actively seeking a new Data Engineer to join their growing team. In this role, the Data Engineer will support the design, automation, and operation of modern data infrastructure.
Responsibilities
Design, build, and maintain scalable and resilient CI/CD pipelines for data applications and infrastructure, with a strong focus on Snowflake, dbt, and modern data tooling
Implement and manage Snowflake dbt projects, including developing dbt models, tests, and documentation, and integrating dbt workflows into CI/CD pipelines
Develop and manage Infrastructure as Code (IaC) using Terraform to provision and configure data infrastructure on Google Cloud Platform (Google Cloud Platform)
Automate the deployment, monitoring, and management of Snowflake data warehouse environments, ensuring optimal performance, security, reliability, and cost efficiency
Collaborate with data engineers and data scientists to understand requirements and deliver automated solutions for data ingestion, transformation, and delivery
Implement and maintain monitoring, logging, and alerting for data pipelines and infrastructure to ensure high availability and proactive issue resolution
Develop and maintain automation scripts and tooling using Python (primary) and Bash for operational tasks
Apply and maintain security best practices across data infrastructure, pipelines, and CI/CD processes
Troubleshoot and resolve issues related to data pipelines, infrastructure, and deployments
Participate in code reviews for Terraform, dbt models, and automation scripts
Create and maintain clear technical documentation for architectures, configurations, and operational processes
Perform other duties, as needed
Qualifications
5+ years of experience in Data Engineering, Analytics Engineering, or DevOps for data platforms
Strong hands-on experience with Snowflake in production environments
Proven expertise with dbt, including model development, testing, and documentation
Experience building and maintaining CI/CD pipelines for data and cloud infrastructure
Hands-on experience with Terraform and Infrastructure as Code practices
Experience working with Google Cloud Platform (Google Cloud Platform) for data storage, processing, and analytics
Strong programming skills in Python and working knowledge of Bash/shell scripting
Experience implementing monitoring, logging, and alerting for data systems
Solid understanding of data pipeline architectures and modern data stack concepts
Strong troubleshooting and problem-solving skills
Ability to collaborate effectively with cross-functional technical teams
Desired Skills
Experience with Data Governance and Data Quality frameworks
Familiarity with Cost Optimization strategies in Snowflake and cloud environments
Experience supporting High-Availability or Mission-Critical Data platforms
Prior contract or consulting experience in fast-paced environments
#J-18808-Ljbffr
Responsibilities
Design, build, and maintain scalable and resilient CI/CD pipelines for data applications and infrastructure, with a strong focus on Snowflake, dbt, and modern data tooling
Implement and manage Snowflake dbt projects, including developing dbt models, tests, and documentation, and integrating dbt workflows into CI/CD pipelines
Develop and manage Infrastructure as Code (IaC) using Terraform to provision and configure data infrastructure on Google Cloud Platform (Google Cloud Platform)
Automate the deployment, monitoring, and management of Snowflake data warehouse environments, ensuring optimal performance, security, reliability, and cost efficiency
Collaborate with data engineers and data scientists to understand requirements and deliver automated solutions for data ingestion, transformation, and delivery
Implement and maintain monitoring, logging, and alerting for data pipelines and infrastructure to ensure high availability and proactive issue resolution
Develop and maintain automation scripts and tooling using Python (primary) and Bash for operational tasks
Apply and maintain security best practices across data infrastructure, pipelines, and CI/CD processes
Troubleshoot and resolve issues related to data pipelines, infrastructure, and deployments
Participate in code reviews for Terraform, dbt models, and automation scripts
Create and maintain clear technical documentation for architectures, configurations, and operational processes
Perform other duties, as needed
Qualifications
5+ years of experience in Data Engineering, Analytics Engineering, or DevOps for data platforms
Strong hands-on experience with Snowflake in production environments
Proven expertise with dbt, including model development, testing, and documentation
Experience building and maintaining CI/CD pipelines for data and cloud infrastructure
Hands-on experience with Terraform and Infrastructure as Code practices
Experience working with Google Cloud Platform (Google Cloud Platform) for data storage, processing, and analytics
Strong programming skills in Python and working knowledge of Bash/shell scripting
Experience implementing monitoring, logging, and alerting for data systems
Solid understanding of data pipeline architectures and modern data stack concepts
Strong troubleshooting and problem-solving skills
Ability to collaborate effectively with cross-functional technical teams
Desired Skills
Experience with Data Governance and Data Quality frameworks
Familiarity with Cost Optimization strategies in Snowflake and cloud environments
Experience supporting High-Availability or Mission-Critical Data platforms
Prior contract or consulting experience in fast-paced environments
#J-18808-Ljbffr