Global Channel Management, Inc
Remote Lead Data Engineer
Global Channel Management, Inc, Dallas, Texas, United States, 75215
Qualifications
Remote Lead Data Engineer needs 7+ years experience as data engineer in consumer finance or equivalent industry (consumer loans, collections, servicing, optional product, and insurance sales).
Strong background in math, statistics, computer science, data science or related discipline.
Advanced knowledge in one of the following languages: Java, Scala, Python, C#.
Production experience with HDFS, YARN, Hive, Spark, Kafka, Oozie / Airflow, Amazon Web Services (AWS), Docker / Kubernetes, Snowflake.
Proficient with:
Data mining/programming tools (e.g., SAS, SQL, R, Python)
Database technologies (e.g., PostgreSQL, Redshift, Snowflake, Greenplum)
Data visualization tools (e.g., Tableau, Looker, MicroStrategy)
Comfortable learning about and deploying new technologies and tools.
Organizational skills and ability to handle multiple projects and priorities simultaneously and meet established deadlines.
Good written and oral communication skills and ability to present results to non-technical audiences.
Knowledge of business intelligence and analytical tools, technologies and techniques.
Additional desired experience:
AWS certification
Spark Streaming
Kafka Streaming / Kafka Connect
ELK Stack
CI/CD: Jenkins, GitLab, Jira, Confluence and other related tools
Responsibilities
Create and manage cloud resources in AWS.
Data ingestion from various data sources (RDBMS, REST HTTP API, flat files, streams, time‑series data) and implement ingestion and processing using Big Data technologies.
Data processing and transformation using technologies such as Spark and cloud services.
Develop automated data quality checks to ensure correct data enters the platform and verify calculation results.
#J-18808-Ljbffr
Strong background in math, statistics, computer science, data science or related discipline.
Advanced knowledge in one of the following languages: Java, Scala, Python, C#.
Production experience with HDFS, YARN, Hive, Spark, Kafka, Oozie / Airflow, Amazon Web Services (AWS), Docker / Kubernetes, Snowflake.
Proficient with:
Data mining/programming tools (e.g., SAS, SQL, R, Python)
Database technologies (e.g., PostgreSQL, Redshift, Snowflake, Greenplum)
Data visualization tools (e.g., Tableau, Looker, MicroStrategy)
Comfortable learning about and deploying new technologies and tools.
Organizational skills and ability to handle multiple projects and priorities simultaneously and meet established deadlines.
Good written and oral communication skills and ability to present results to non-technical audiences.
Knowledge of business intelligence and analytical tools, technologies and techniques.
Additional desired experience:
AWS certification
Spark Streaming
Kafka Streaming / Kafka Connect
ELK Stack
CI/CD: Jenkins, GitLab, Jira, Confluence and other related tools
Responsibilities
Create and manage cloud resources in AWS.
Data ingestion from various data sources (RDBMS, REST HTTP API, flat files, streams, time‑series data) and implement ingestion and processing using Big Data technologies.
Data processing and transformation using technologies such as Spark and cloud services.
Develop automated data quality checks to ensure correct data enters the platform and verify calculation results.
#J-18808-Ljbffr