Bachelor's degree in a relevant technical discipline or equivalent experience required., Expert level skills in ETL/Data Engineering Solution Design and Architecture, SQL, and Data Modeling., 5+ years of relevant experience in data engineering and team leadership capabilities., Familiarity with modern data technologies such as Databricks, Python, and MSBI is essential..
Key responsabilities:
Design and automate distributed systems for data ingestion and transformation.
Implement frameworks to monitor and troubleshoot data quality and integrity issues.
Build reliable and scalable data pipelines using ETL/ELT tools and scripting languages.
Coach and develop less experienced team members while ensuring successful analytics initiatives.
Report This Job
Help us maintain the quality of our job listings. If you find any issues with this job post, please let us know.
Select the reason you're reporting this job:
At Cummins, we empower everyone to grow their careers through meaningful work, building inclusive and equitable teams, coaching, development and opportunities to make a difference. Across our entire organization, you'll find engineers, developers, and technicians who are innovating, designing, testing, and building. You'll also find accountants, marketers, as well as manufacturing, quality and supply chain specialists who are working with technology that's just as innovative and advanced.
From your first day at Cummins, weβre focused on understanding your talents, current skills and future goals β and creating a plan to get you there. Your journey begins with planning your development and connecting to diverse experiences designed to spur innovation. From our internships to our senior leadership roles, we attract, hire and reward the best and brightest from around the world and look to them for new ideas and fresh perspectives. Learn more about #LifeAtCummins at cummins.com/careers.
Although the role category specified in the GPP is Remote, the requirement is for Hybrid.
Key Responsibilities
Design and Automation : Deploy distributed systems for ingesting and transforming data from various sources (relational, event-based, unstructured).
Data Quality and Integrity : Implement frameworks to monitor and troubleshoot data quality and integrity issues.
Data Governance : Establish processes for managing metadata, access, and retention for internal and external users.
Data Pipelines : Build reliable, efficient, scalable, and quality data pipelines with monitoring and alert mechanisms using ETL/ELT tools or scripting languages.
Database Structure : Design and implement physical data models to optimize database performance through efficient indexing and table relationships.
Optimization and Troubleshooting : Optimize, test, and troubleshoot data pipelines.
Large Scale Solutions : Develop and operate large-scale data storage and processing solutions using distributed and cloud-based platforms (e.g., Data Lakes, Hadoop, Hbase, Cassandra, MongoDB, Accumulo, DynamoDB).
Automation : Use modern tools and techniques to automate common, repeatable, and tedious data preparation and integration tasks.
Infrastructure Renovation : Renovate data management infrastructure to drive automation in data integration and management.
Agile Development : Ensure the success of critical analytics initiatives using agile development technologies such as DevOps, Scrum, Kanban.
Team Development : Coach and develop less experienced team members.
Responsibilities
Qualifications:
College, university, or equivalent degree in a relevant technical discipline, or equivalent experience required. Licensing may be required for compliance with export controls or sanctions regulations.
Competencies
System Requirements Engineering : Translate stakeholder needs into verifiable requirements; establish acceptance criteria; track requirements status; assess impact of changes.
Collaboration : Build partnerships and work collaboratively to meet shared objectives.
Communication : Develop and deliver communications that convey a clear understanding of the unique needs of different audiences.
Decision Quality : Make good and timely decisions to keep the organization moving forward.
Data Extraction : Perform ETL activities from various sources using appropriate tools and technologies.
Programming : Create, write, and test computer code, test scripts, and build scripts to meet business, technical, security, governance, and compliance requirements.
Quality Assurance Metrics : Apply measurement science to assess solution outcomes using ITOM, SDLC standards, tools, metrics, and KPIs.
Solution Documentation : Document information and solutions to enable improved productivity and effective knowledge transfer.
Solution Validation Testing : Validate configuration item changes or solutions using SDLC standards, tools, and metrics.
Data Quality : Identify, understand, and correct data flaws to support effective information governance.
Problem Solving : Solve problems using systematic analysis processes; implement robust, data-based solutions; prevent problem recurrence.
Values Differences : Recognize the value of different perspectives and cultures.
Qualifications
Skills:
ETL/Data Engineering Solution Design and Architecture : Expert level.
SQL and Data Modeling : Expert level (ER Modeling and Dimensional Modeling).
Team Leadership : Ability to lead a team of data engineers.
MSBI (SSIS, SSAS) : Experience required.
Databricks (Pyspark) and Python : Experience required.
Additional Skills : Snowflake, Power BI, Neo4j (good to have).
Communication : Good communication skills.
Preferred Experience
8+ years of overall experience.
5+ years of relevant experience in data engineering.
Knowledge of the latest technologies and trends in data engineering.
Technologies : Familiarity with analyzing complex business systems, industry requirements, and data regulations.
Big Data Platform : Design and development using open source and third-party tools.