Siri InfoSolutions Inc
Overview
Data Engineer – Location: Plano, TX – Duration: Long Term Logistics: Plano based roles; normal expectations – 3 days in the office Project & Domain: Enterprise productivity; this team has built a platform for their internal teams to leverage AI. Difference from before: previously mainly POC localizing parts of Cap One; now will scale across all of enterprise. Responsibilities
Build data pipelines and provide ongoing support (pipeline development and production support). Contribute to data platform efforts that enable enterprise-wide AI tooling. Technical Stack
SQL TypeScript or Python (for pipeline development) Kafka Snowflake Core Data Engineering Skills
Experience in building data pipelines and app development including Python, SQL, TypeScript Experience with AWS and distributed tools (Kafka - preference optional - MapReduce Hadoop Hive EMR Spark; Gurobi or MySQL) Experience with ETL/ELT tools (Airbyte - preference Airflow, etc.) Experience working on real-time data and streaming applications (Kafka, Flink, etc.) Experience with data warehousing (Snowflake) Experience with UNIX/Linux including basic commands and shell scripting Experience with Agile engineering practices Data Analysis Skills
SQL: understanding of tables/views beyond utility for consumption Ability to analyze differences between source systems and platform representations Ability to analyze data produced by Snowflake Able to work in ambiguous or "gray" areas Key Skills
Apache Hive, S3, Hadoop, Redshift, Spark, AWS, Apache Pig, NoSQL, Big Data, Data Warehouse, Kafka, Scala Employment Type :
Full Time Experience:
years Vacancy:
1
#J-18808-Ljbffr
Data Engineer – Location: Plano, TX – Duration: Long Term Logistics: Plano based roles; normal expectations – 3 days in the office Project & Domain: Enterprise productivity; this team has built a platform for their internal teams to leverage AI. Difference from before: previously mainly POC localizing parts of Cap One; now will scale across all of enterprise. Responsibilities
Build data pipelines and provide ongoing support (pipeline development and production support). Contribute to data platform efforts that enable enterprise-wide AI tooling. Technical Stack
SQL TypeScript or Python (for pipeline development) Kafka Snowflake Core Data Engineering Skills
Experience in building data pipelines and app development including Python, SQL, TypeScript Experience with AWS and distributed tools (Kafka - preference optional - MapReduce Hadoop Hive EMR Spark; Gurobi or MySQL) Experience with ETL/ELT tools (Airbyte - preference Airflow, etc.) Experience working on real-time data and streaming applications (Kafka, Flink, etc.) Experience with data warehousing (Snowflake) Experience with UNIX/Linux including basic commands and shell scripting Experience with Agile engineering practices Data Analysis Skills
SQL: understanding of tables/views beyond utility for consumption Ability to analyze differences between source systems and platform representations Ability to analyze data produced by Snowflake Able to work in ambiguous or "gray" areas Key Skills
Apache Hive, S3, Hadoop, Redshift, Spark, AWS, Apache Pig, NoSQL, Big Data, Data Warehouse, Kafka, Scala Employment Type :
Full Time Experience:
years Vacancy:
1
#J-18808-Ljbffr