Strong experience in ETL optimization and designing big data processes using Apache Spark or similar technologies., Proficiency in building scalable data pipelines with SQL, Python, Spark, or PySpark, with advanced knowledge in at least one programming language., Experience maintaining and enhancing Confluent Kafka architecture, including design principles and CI/CD deployment procedures., Knowledge of real-time streaming applications, Kafka producers, consumers, and streams, with a background in distributed systems and data architecture..
Key responsibilities:
Lead the development and optimization of data pipelines and streaming applications using Kafka and Spark.
Maintain and enhance Kafka architecture, ensuring adherence to design principles and deployment procedures.
Provide technical leadership and mentorship to junior engineers in data engineering best practices.
Collaborate in a fast-paced, agile environment to deliver scalable, real-time data processing solutions.
Report this Job
Help us maintain the quality of our job listings. If you find any issues
with this job post, please let us know. Select the reason you're reporting
this job:
Coders Brain is a global leader in IT services, digital and business solutions that partners with its clients to simplify, strengthen and transform their businesses. We ensure the highest levels of certainty and satisfaction through a deep-set commitment to our clients, comprehensive industry expertise and a global network of innovation and delivery centers.
We achieved our success because of how successfully we integrate with our clients.
Expertise in ETL optimization, designing, coding, and tuning big data processes using Apache Spark or similar technologies.
Experience building robust and scalable data integration (ETL) pipelines using SQL, Python, Spark or PySpark. Advanced knowledge in one of the programming language is must.
Experience with building streaming applications with Confluent Kafka (Confluent Kafka preferred but opensource Kafka acceptable)
Development experience using Kafka producers, consumers, and streams (Confluent Kafka preferred but opensource Kafka acceptable)
Experience with building data pipelines and applications to stream and process datasets at low latencies.
Experience with realtime and scalable systems development using Apache Kafka or Confluent Kafka or Kafka Streams.
Show efficiency in handling data tracking data lineage, ensuring data quality, and improving discoverability of data.
Good understanding of AWS technologies (S3, AWS Glue, CDK, ECS, EMR, Redshift, Athena)
Sound knowledge of distributed systems and data architecture (lambda) design and implement batch and stream data processing pipelines, knows how to optimize the distribution, partitioning, and MPP of highlevel data structures.
Knowledge of Engineering and Operational Excellence using standard methodologies.
Experience with process improvement, workflow, bench marking and or evaluation of business processes.
Familiarity with CICD process.
Work in a fast paced agile environment.
Experience providing technical leadership and mentoring other junior engineers for best practices on data engineering.
Experience in building RESTAPIs for data transfers.
Background in Java and Spring framework is a plus
Proficiency in one of JIRA, Atlassian, and Git is must.
Requirements
Experience with realtime and scalable systems development using Apache Kafka or Confluent Kafka or Kafka Streams.
Experience providing technical leadership and mentoring other junior engineers for best practices on data engineering.
Experience in building RESTAPIs for data transfers.
Background in Java and Spring framework is a plus
Proficiency in one of JIRA, Atlassian, and Git is must.
Required profile
Experience
Level of experience:Senior (5-10 years)
Spoken language(s):
English
Check out the description to know which languages are mandatory.