r2 Technologies, Inc.
Position: Data Architect
Location: Hartford, CT (Remote right now)
Duration: 12+Months
Mandatory Skills
: Talend, PySpark, Redshift
Job Description
Responsibilities:
Translating data and technology requirements into our ETL / ELT architecture.
Develop real-time and batch data ingestion and stream-analytic solutions leveraging technologies such as Kafka, Apache Spark, Java, NoSQL DBs, AWS EMR.
Develop data driven solutions utilizing current and next generation technologies to meet evolving business needs.
Develop custom cloud-based data pipeline.
Provide support for deployed data applications and analytical models by identifying data problems and guiding issue resolution with partner data engineers and source data providers.
Provide subject matter expertise in the analysis, preparation of specifications and plans for the development of data processes.
Qualifications:
Strong experience in data ingestion, gathering, wrangling and cleansing tools such as Apache NiFI, Kylo, Scripting, Power BI, Tableau and/or Qlik
Experience with data modeling, data architecture design and leveraging large-scale data ingest from complex data sources
Experience building and optimizing 'big data' data pipelines, architectures and data sets.
Advanced SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
Strong knowledge of analysis tools such as Python, R, Spark or SAS, Shell scripting, R/Spark on Hadoop or Cassandra preferred.
Strong knowledge of data pipelining software e.g., Talend, Informatica
Skills:
Talend,PySpark,Amazon Redshift,ETL,ELT,Kafka,Apache Spark,Java,NoSQL,AWS EMR.,data pipeline,Apache NiFI,Kylo,Scripting,Power BI,Tableau,Qlik,data ingestion,wrangling,cleansing,Advanced SQL,SQL,Python,Hadoop,Cassandra,R/Spark,Informatica
Location: Hartford, CT (Remote right now)
Duration: 12+Months
Mandatory Skills
: Talend, PySpark, Redshift
Job Description
Responsibilities:
Translating data and technology requirements into our ETL / ELT architecture.
Develop real-time and batch data ingestion and stream-analytic solutions leveraging technologies such as Kafka, Apache Spark, Java, NoSQL DBs, AWS EMR.
Develop data driven solutions utilizing current and next generation technologies to meet evolving business needs.
Develop custom cloud-based data pipeline.
Provide support for deployed data applications and analytical models by identifying data problems and guiding issue resolution with partner data engineers and source data providers.
Provide subject matter expertise in the analysis, preparation of specifications and plans for the development of data processes.
Qualifications:
Strong experience in data ingestion, gathering, wrangling and cleansing tools such as Apache NiFI, Kylo, Scripting, Power BI, Tableau and/or Qlik
Experience with data modeling, data architecture design and leveraging large-scale data ingest from complex data sources
Experience building and optimizing 'big data' data pipelines, architectures and data sets.
Advanced SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
Strong knowledge of analysis tools such as Python, R, Spark or SAS, Shell scripting, R/Spark on Hadoop or Cassandra preferred.
Strong knowledge of data pipelining software e.g., Talend, Informatica
Skills:
Talend,PySpark,Amazon Redshift,ETL,ELT,Kafka,Apache Spark,Java,NoSQL,AWS EMR.,data pipeline,Apache NiFI,Kylo,Scripting,Power BI,Tableau,Qlik,data ingestion,wrangling,cleansing,Advanced SQL,SQL,Python,Hadoop,Cassandra,R/Spark,Informatica