Logo for Keylent Inc

Spark Developer

Roles & Responsibilities

  • Experience in developing Spark applications using Spark RDD API, Spark SQL, GraphX, Streaming API, MLlib, DataFrames, and Spark on Yarn.
  • Strong knowledge of Spark performance tuning, workflows, how to write Spark jobs, and query optimization.
  • Solid foundation in data structures and algorithms with hands-on experience in at least one programming language (Scala preferred, Python second, Java third).
  • Experience with data ingestion, transformation, optimization techniques, and aggregation pipeline design; familiarity with big data storage systems and tools (Hadoop/HDFS, AWS S3, Kafka, Elasticsearch, Cassandra, MongoDB, DynamoDB, PostgreSQL/MySQL).

Requirements:

  • Create new, and maintain existing Scala/Spark jobs for data transformation and aggregation from simple to complex data transformations involving structured and unstructured data.
  • Produce unit tests for Spark transformations and helper methods.
  • Develop data processing pipelines, data storage, and management architecture.
  • Define scalable calculation logic for interactive and batch use cases.

Job description


Role: Spark Developer

Client: Client

Location: Remote

Experience in developing Spark Applications using Spark RDD API, Spark-SQL, Spark GraphX API, Spark Streaming API, Spark-Yarn, Spark MLib API, and Data frame APIs
Should have a broad knowledge of Spark Advantages, Spark Workflows, How to write Spark Jobs, Spark query tuning, and performance optimization.
Solid in Data Structures, Algorithms basics and Should have good hands-on experience with any - one programming language (Scala/ Java8 - 1st Preference OR Python - 2nd Preference OR Java - 3rd preference) and Strong investigative and problem-solving skills
Data Ingestion, Optimization Techniques, Data Transformation, and aggregation pipeline design/development knowledge is required.
Experience working on cutting-edge Big Data storage systems and technologies like Hadoop, HDFS, AWS S3, AWS Lambda, Storm/Heron, Cassandra, Apache Kafka, Solr/ElasticSearch, MongoDB, DynamoDB, Postgres, and/or MySQL, etc.

Roles & Responsibilities:
Create new, and maintain existing Scala/Spark jobs for data transformation and aggregation from simple to Complex Data transformations involving structured & unstructured data.
Produce unit tests for Spark transformations and helper methods
Develop data processing pipelines, data storage, and management architecture.
Define scalable calculation logic for interactive and batch use cases
Interact with infrastructure and data teams to produce complex analyses across data
You'll be working on a unique and challenging big data ecosystem with a focus on storage efficiency, data security, privacy, scalable and performant queries, expandability and flexibility, etc, with the goal to help better measure the quality of map data.
You will work with engineers to build a big data platform that processes and manages Exabytes of data and enables efficient access to those


Related jobs

Other jobs at Keylent Inc

We help you get seen. Not ignored.

We help you get seen faster — by the right people.

🚀

Auto-Apply

We apply for you — automatically and instantly.

Save time, skip forms, and stay on top of every opportunity. Because you can't get seen if you're not in the race.

✨

AI Match Feedback

Know your real match before you apply.

Get a detailed AI assessment of your profile against each job posting. Because getting seen starts with passing the filters.

Upgrade to Premium. Apply smarter and get noticed.

Upgrade to Premium

Join thousands of professionals who got noticed and hired faster.