Data Analyst / Engineer Columbus, OH W2 Only Full AWS Environment Good understanding about Teradata, Glue Data Migration JD : Responsibilities: Develop and maintain data platforms using Python, Spark, and PySparknbsp; Handle migration to PySpark on AWSnbsp; Design and implement data pipelinesnbsp; Work with AWS and Big Datanbsp; Produce unit tests for Spark transformations and helper methodsnbsp; Create Scala/Spark jobs for data transformation and aggregationnbsp; Write Scaladoc-style documentation for codenbsp; Optimize Spark queries for performancenbsp; Integrate with SQL databases (e.g., Microsoft, Oracle, Postgres, MySQL Understand distributed systems concepts (CAP theorem, partitioning, replication, consistency, and consensus Skills: Proficiency in Python, Scala (with a focus on functional programming), and Sparknbsp; Familiarity with Spark APIs, including RDD, DataFrame, MLlib, GraphX, and Streamingnbsp; Experience working with HDFS, S3, Cassandra, and/or DynamoDBnbsp; Deep understanding of distributed systemsnbsp; Experience with building or maintaining cloud-native applicationsnbsp; Familiarity with serverless approaches using AWS Lambda is a plus Note: This is aW2Contract.
So, candidate must work onGalaxy I TechPayroll.
For Immediate response please reach out to me atsrinivas at galaxyitech dot com / Four Eight Zero - Four Zero Seven - Six Nine Three Zero.