Logo
HashRoot Limited

Snowflake Data Engineer

HashRoot Limited, Dallas, Texas, United States, 75215

Save Job

We are seeking an experienced Snowflake Data Engineer to design and optimize Snowflake-based data solutions, build real-time and batch ingestion pipelines, implement advanced features like Streams, Tasks, and Snowpark, and ensure secure, scalable, and cost-effective data platforms. The role involves close collaboration with architects and domain teams to deliver data products and enable integration with BI, governance, and cloud platforms. Key Responsibilities

Design and implement Snowflake schemas (star, snowflake, data vault) optimized with micro-partitioning, clustering keys, materialized views, and search optimization services. Build real-time and batch ingestion pipelines into Snowflake using Snowpipe, Kafka Connect, Fivetran, Matillion, Informatica, or dbt. Automate incremental data processing with Streams & Tasks to support CDC (Change Data Capture). Use Zero-Copy Cloning for environment management, testing, and sandboxing. Apply Time Travel and Fail-safe features for data recovery and auditing. Develop data transformation logic in Snowpark for Python/SQL/Scala to push compute directly into Snowflake. Design integrations with cloud storage (S3, Azure ADLS, GCS) for staging and external tables. Implement data sharing and data marketplace solutions via Snowflake Secure Data Sharing and Snowflake Marketplace. Enable semi-structured data handling (JSON, Avro, Parquet, ORC, XML) using VARIANT columns and lateral flattening. Integrate Snowflake with BI tools (Power BI, Tableau) via live connections and semantic layers. Implement RBAC (Role-Based Access Control), Row Access Policies, and Dynamic Data Masking for data security. Optimize compute usage with multi-cluster warehouses, resource monitors, and query performance tuning. Manage cost optimization strategies (warehouse auto-suspend, query profiling, storage/compute separation). Integrate with data catalog & governance platforms (Collibra, Alation, Informatica CDGC) using Snowflake metadata and APIs. Work with domain teams to deliver data products leveraging Snowflakes data mesh-friendly features. Collaborate with architects to design a Snowflake-centric data fabric integrated with ETL/ELT and API layers. Support CI/CD automation for Snowflake code deployment using GitHub Actions, Azure DevOps, or dbt Cloud. Qualifications

Bachelors or Masters in Computer Science, Data Engineering, or related field. 10+ years of data engineering experience, with 5+ years in Snowflake Data Cloud. Expertise in SQL optimization and Snowflake performance tuning. Hands-on with Snowpipe, Streams & Tasks, Snowpark, Zero-Copy Cloning, and Secure Data Sharing. Proficiency in Python, Scala, or Java for Snowpark development. Experience integrating with cloud platforms like AWS. Exposure to ETL/ELT tools (Informatica, Matillion, Fivetran). Familiarity with CI/CD, Git, DevOps practices for data operations. Preferred Certifications

Not Applicable #J-18808-Ljbffr