Spark Developer_Quest Global Pvt Ltd

Remote: 
Full Remote
Contract: 
Work from: 

Offer summary

Qualifications:

Bachelor's or Master's degree in Computer Science or related field (BE/BTech/MCA/MTech/MSC), 3-5 years of experience in Apache Spark development or related Data Engineering frameworks, Proficiency in Java/Scala programming, with Python as a plus, Experience with Big Data technologies and tools like Kafka, NoSQL databases, and ETL processes..

Key responsibilities:

  • Develop scalable distributed architecture as defined by the Architect or Tech Lead.
  • Analyze and assemble large data sets to meet functional and non-functional requirements.
  • Create and optimize ETL scripts for big data sources and automate reporting processes.
  • Collaborate with stakeholders to ensure all end-to-end requirements are met and adhere to the Software Development Life Cycle (SDLC).

CodersBrain logo
CodersBrain SME https://www.codersbrain.com/
201 - 500 Employees
See all jobs

Job description

Experience:3 to 5
  
Responsibilities 
 .   As a Spark developer you will manage the development of scalable distributed Architecture
defined by the Architect or tech Lead in our team. 
 .   Analyse, assemble large data sets to designed for the functional and non-functional
requirements. 
 .   You will develop ETL scripts for big data sources. 
 .   Identify, design & optimise data processing & automate for reports and dashboards. 
 .   You will be responsible for workflow optimizations, data optimizations and ETL optimization
as per the requirements elucidated by the team. 
 .   Work with stakeholders such as Product managers, Technical Leads & Service Layer
engineers to ensure end-to-end requirements are addressed. 
 .   Strong team player to adhere to Software Development Life cycle (SDLC) and
documentations needed to represent every stage of SDLC. 
 General Qualifications: 
 .   BE/BTech/MCA/MTech/MSC with 3-5 years of experience with Apache Spark development
or in alternative Data Engineering & analytics frameworks. 
 .   Overall experience in the field Java/Scala as a Backend developer could be between 3 years. 
 .   Working experience in Bigdata is important. 
 Technical Skills 
 .   Programming experience in Java/Scala is needed. Experience in Python also will have an
advantage. 
 .   Writing ETL Stack in scalable & optimized fashion using Apache Spark/Hadoop/Kafka etc. 
 .   Should have working experience in writing distributed & optimized Apache Spark jobs for
various Machine Learning Algorithms. 
 .   Experience building and optimizing data pipeline and data sets is essential. 
 .   Should have worked with Kafka, at least one of No SQL databases (e.g. Cassandra,
MongoDB, elastic DB) and at least one of RDBMS (e.g. MySQL) 
 .   Working experience to use cache such as Redis, Apache Ignite, Hazelcast will advantage. 
 .   Working knowledge of Docker and Kubernetes will be a plus. 
Kindly find the below inputs from customer to get quality profiles
 Include the criteria apart from Spark à Experience in No sql databases (elastic search OR
Cassandra OR MongoDB), Apache Kafka.
 Look for Spark Streaming (people who have used Sqoop, likely doesn’t know Spark Streaming).
 Scala or Java based work experience using Spark is a MUST. Python usage with Spark is
optional, not must.

Spark streaming ,Kafka experience is mandatory for this requirement.

Required profile

Experience

Spoken language(s):
English
Check out the description to know which languages are mandatory.

Other Skills

  • Teamwork
  • Communication
  • Problem Solving

Related jobs