Requirements
Key Responsibilities
· Design and implement data products and features in collaboration with product owners, data analysts, and business partners using Agile / Scrum methodology
· Contribute to overall architecture, frameworks and patterns for processing and storing large data volumes
· Translate product backlog items into engineering designs and logical units of work
· Profile and analyze data for the purpose of designing scalable solutions
· Define and apply appropriate data acquisition and consumption strategies for given technical scenarios
· Design and implement distributed data processing pipelines using tools and languages prevalent in the big data ecosystem
· Build utilities, user defined functions, libraries, and frameworks to better enable data flow patterns
· Implement complex automated routines using workflow orchestration tools
· Work with architecture, engineering leads and other teams to ensure quality solutions are implemented, and engineering best practices are defined and adhered to
· Anticipate, identify and solve issues concerning data management to improve data quality
· Build and incorporate automated unit tests and participate in integration testing efforts
· Utilize and advance continuous integration and deployment frameworks
· Troubleshoot data issues and perform root cause analysis
· Work across teams to resolve operational & performance issues
Qualifications
· MS/BS in Computer Science, or related technical discipline
· 4+ years of experience in large-scale software development, 2+ years of big data experience
· Strong programming experience, Python preferred
· Extensive experience working with Hadoop and related processing frameworks such as Spark, Hive, Sqoop, etc.
· Experience with RDBMS systems, SQL and SQL Analytical functions
· Experience with workflow orchestration tools like Apache Airflow
· Experience with performance and scalability tuning
Nice to have
· Experience with Scala or Java
· Experience working in a public cloud environment, particularly AWS
· Familiarity with cloud warehouse tools like Snowflake
· Experience with messaging/streaming/complex event processing tooling and frameworks such as Kinesis, Kafka, Spark Streaming, Flink, Nifi, etc.
· Experience working with NoSQL data stores such as HBase, DynamoDB, etc.
· Experience building RESTful API’s to enable data consumption
· Familiarity with build tools such as Terraform or CloudFormation and automation tools such as Jenkins or Circle CI
· Familiarity with practices like Continuous Development, Continuous Integration and Automated Testing
· Experience in Agile/Scrum application development
· Familiarity with practices like Continuous Development, Continuous Integration and Automated Testing
· Experience in Agile/Scrum application development
· Desire to work collaboratively with your teammates to come up with the best solution to a problem
· Demonstrated experience and ability to deliver results on multiple projects in a fast-paced, agile environment
· Excellent problem-solving and interpersonal communication skills
· Strong desire to learn and share knowledge with others
· Desire to work collaboratively with your teammates to come up with the best solution to a problem
· Demonstrated experience and ability to deliver results on multiple projects in a fast-paced, agile environment
· Excellent problem-solving and interpersonal communication skills
Strong desire to learn and share knowledge with other