Galaxy i Technologies, Inc
Data Analyst / Engineer
Columbus, OH
W2 Only
Full AWS Environment
Good understanding about Teradata, Glue
Data Migration JD :
Responsibilities:
Develop and maintain data platforms using Python, Spark, and PySparknbsp;
Handle migration to PySpark on AWSnbsp;
Design and implement data pipelinesnbsp;
Work with AWS and Big Datanbsp;
Produce unit tests for Spark transformations and helper methodsnbsp;
Create Scala/Spark jobs for data transformation and aggregationnbsp;
Write Scaladoc-style documentation for codenbsp;
Optimize Spark queries for performancenbsp;
Integrate with SQL databases (e.g., Microsoft, Oracle, Postgres, MySQL
Understand distributed systems concepts (CAP theorem, partitioning, replication, consistency, and consensus
Skills:
Proficiency in Python, Scala (with a focus on functional programming), and Sparknbsp; Familiarity with Spark APIs, including RDD, DataFrame, MLlib, GraphX, and Streamingnbsp; Experience working with HDFS, S3, Cassandra, and/or DynamoDBnbsp; Deep understanding of distributed systemsnbsp; Experience with building or maintaining cloud-native applicationsnbsp; Familiarity with serverless approaches using AWS Lambda is a plus
Note: This is aW2Contract. So, candidate must work onGalaxy I TechPayroll. For Immediate response please reach out to me atsrinivas at galaxyitech dot com / Four Eight Zero - Four Zero Seven - Six Nine Three Zero.
Columbus, OH
W2 Only
Full AWS Environment
Good understanding about Teradata, Glue
Data Migration JD :
Responsibilities:
Develop and maintain data platforms using Python, Spark, and PySparknbsp;
Handle migration to PySpark on AWSnbsp;
Design and implement data pipelinesnbsp;
Work with AWS and Big Datanbsp;
Produce unit tests for Spark transformations and helper methodsnbsp;
Create Scala/Spark jobs for data transformation and aggregationnbsp;
Write Scaladoc-style documentation for codenbsp;
Optimize Spark queries for performancenbsp;
Integrate with SQL databases (e.g., Microsoft, Oracle, Postgres, MySQL
Understand distributed systems concepts (CAP theorem, partitioning, replication, consistency, and consensus
Skills:
Proficiency in Python, Scala (with a focus on functional programming), and Sparknbsp; Familiarity with Spark APIs, including RDD, DataFrame, MLlib, GraphX, and Streamingnbsp; Experience working with HDFS, S3, Cassandra, and/or DynamoDBnbsp; Deep understanding of distributed systemsnbsp; Experience with building or maintaining cloud-native applicationsnbsp; Familiarity with serverless approaches using AWS Lambda is a plus
Note: This is aW2Contract. So, candidate must work onGalaxy I TechPayroll. For Immediate response please reach out to me atsrinivas at galaxyitech dot com / Four Eight Zero - Four Zero Seven - Six Nine Three Zero.