Hands-on PySpark development with Python/Scala (preferred) or Java/Scala
Experience with Core Java, MapReduce, Hive programming and Hive performance concepts
Experience with Git repositories and source code management
Requirements:
Design and develop Hadoop-based applications and PySpark jobs
Implement MapReduce, Hive queries and optimize performance
Collaborate in AWS ecosystem (EC2, S3) and maintain CI/CD workflows using Git and Jenkins
Participate in Agile software development and deliver solutions iteratively
Job description
This is a remote position.
Primary Skills ● Design and develop on Hadoop applications ● Hands-on in developing Jobs in pySpark with Python/ SCALA (Preferred) or Java/ SCALA ● Experience on Core Java, Experience on Map Reduce programs, Hive programming, Hive queries performance concepts ● Experience on source code management with Git repositories
Secondary Skills ● Exposure to AWS Ecosystem with hands-on knowledge of ec2, S3 and services ● Basic SQL programming ● Knowledge of agile methodology for delivering software solutions ● Build scripting with Maven / Cradle, Exposure to Jenkins