Help us maintain the quality of our job listings. If you find any issues with this job post, please let us know.
Select the reason you're reporting this job:
Datavail is a proven technology consulting, implementation, and managed services partner headquartered in Boulder, Colorado. Our world-class tech-enabled delivery platform enables us to provide an extensive solutions portfolio, including database management, analytics, application development, cloud infrastructure, and enterprise applications consulting — on-premises and in the cloud. With tens of thousands of hours of experience, our goal is to enable our clients to build modern, data-driven organizations that scale.
Datavail maintains strong partnerships with leading technology companies to help our clients leverage the best-fit platform to optimize outcomes. In addition, we have poured over $25 million in R&D to bring world-class systems and application monitoring tools to the market. We seek to leverage the greatest that technology and talent have to offer to simplify and operationalize the IT systems organizations use to manage their most valuable data.
We select the best across the globe so we can implement complex solutions for top brands as well as small-to-midsize organizations. Our passionate team of 1200+ certified, cloud-focused consultants deliver industry-focused solutions to 500+ customers with an average lifespan of 7+ years. These committed tech professionals help our clients jump-start their projects, improve time to value & reduce the risk for the core industries that we serve.
Our Technology Alliances:
Datavail has leading partnerships with AWS, Microsoft, Oracle, Google, Maria DB, Mongo DB, MySQL and can provide 24×7 managed services for all leading database and applications environments including:
· AWS
· Azure
· Oracle
· MongoDB
· SQL Server
· Db2
· PostgreSQL
· MySQL
· Cassandra
· Hadoop
· MariaDB
· Oracle E-Business Suite (EBS)/ Oracle PeopleSoft/ Oracle Hyperion technologies
· Oracle Enterprise Resource Planning (ERP) and Enterprise Performance Management (EPM) technologies
· Oracle Analytics & Business Intelligence tech.
Data Engineer is responsible for designing, building, and maintaining the infrastructure and systems required for collecting, storing, and processing large datasets efficiently.
Education: Bachelor's degree in computer science with 8+ years of experience
Experience:
Technical Skills
Programming Languages: Proficiency in Python, SQL, Java, or Scala for data manipulation and pipeline development.
Data Processing Frameworks: Experience with tools like Apache Spark, Hadoop, or Apache Kafka for large-scale data processing.
Data Systems and Platforms
Databases: Knowledge of both relational databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra).
Data Warehousing: Experience with platforms like Snowflake, Amazon Redshift and Azure Synapse.
Cloud Platforms: Familiarity with AWS, Azure Cloud for deploying and managing data pipelines. Having Good experience in Fabric is advantageous
Experience working with distributed computing systems such as Hadoop HDFS, Hive, or Spark.
Managing and optimizing data lakes and delta lakes for structured and unstructured data.
Data Modeling and Architecture
Expertise in designing efficient data models (e.g., star schema, snowflake schema) and maintaining data integrity.
Understanding of modern data architectures like Data Mesh or Lambda Architecture.
Data Pipeline Development
Building and automating ETL/ELT pipelines for extracting data from diverse sources, transforming it, and loading it into target systems.
Monitoring and troubleshooting pipeline performance and failures.
Workflow Orchestration
Hands-on experience with orchestration tools such as Azure Data Factory, AWS Glue jobs, DMS or Prefect to schedule and manage workflows.
Version Control and CI/CD
Utilizing Git for version control and implementing CI/CD practices for data pipeline deployments.
Key Skills:
Proficiency in programming languages such as Python, SQL, and optionally Scala or Java.
Proficiency in data processing frameworks like Apache Spark and Hadoop is crucial for handling large-scale and real-time data.
Expertise in ETL/ELT tools such as Azure ADF and Fabric Data Pipeline is important for creating efficient and scalable data pipelines.
A solid understanding of database systems, including relational databases like MySQL and PostgreSQL, as well as NoSQL solutions such as MongoDB and Cassandra, is fundamental.
Experience with cloud platforms, including AWS, Azure and their data-specific services like S3, BigQuery, and Azure Data Factory, is highly valuable.
Data modeling skills, including designing star or snowflake schema, and knowledge of modern architectures like Lambda and Data Mesh, are critical for building scalable solutions.
Role and Responsibilities:
Responsible for designing, developing, and maintaining data pipelines and infrastructure to support our data-driven decision-making processes.
Design, build, and maintain data pipelines to extract, transform, and load data from various sources into our data warehouse and data lake.
Proficient in creating data bricks creating notebooks, working with catalogs, native SQL, creating clusters, Parameterizing notebooks, and administrating data bricks. Define security models and assign roles as per requirement.
Responsible for creating data flow in Synapse analytics integrating external source systems, creating external tables, data flows and create data models. Schedule the pipelines using various jobs, creating trigger
Design and develop data pipelines using Fabric pipelines, spark notebooks accessing multiple data sources. Proficient in developing Data bricks notebooks and data optimization
Develop and implement data models to ensure data integrity and consistency. Manage and optimize data storage solutions, including databases and data warehouses.
Develop and implement data quality checks and validation procedures to ensure data accuracy and reliability.
Design and implement data infrastructure components, including data pipelines, data lakes, and data warehouses.
Collaborate with data scientists, analysts, and other stakeholders to understand business requirements and translate them into technical solutions.
Monitoring Azure and Fabric data pipelines, spark jobs and work on fixes based on the request priority.
Responsible for data monitoring activities, having good knowledge on reporting tools like Power Bi and Tableau is required.
Responsible for understanding the client requirements and architect solutions in both Azure and AWS cloud platforms.
Monitor and optimize data pipeline performance and scalability to ensure efficient data processing.
About Us
Datavail is a leading provider of data management, application development, analytics, and cloud services, with more than 1,000 professionals helping clients build and manage applications and data via a world-class tech-enabled delivery platform and software solutions across all leading technologies. For more than 17 years, Datavail has worked with thousands of companies spanning different industries and sizes, and is an AWS Advanced Tier Consulting Partner, a Microsoft Solutions Partner for Data & AI and Digital & App Innovation (Azure), an Oracle Partner, and a MySQL Partner.
About The Team
Datavail’s Data Management Services:
Datavail’s Data Management and Analytics practice is made up of experts who provide a variety of data services including initial consulting and development, designing and building complete data systems, as well as ongoing support and management of database, data warehouse, data lake, data integration, and virtualization and reporting environments. Datavail’s team is comprised of not just excellent BI & analytics consultants, but great people as well. Datavail’s data intelligence consultants are experienced, knowledgeable and certified in the best in breed BI and analytics software applications and technologies. We ascertain your business objectives, goals and requirements, assess your environment, and recommend the tools which best fit your unique situation. Our proven methodology can help your project succeed, regardless of stage. With the combination of a proven delivery model and top-notch experience ensures that Datavail will remain the Data Management experts on demand you desire. Datavail’s flexible and client focused services always add value to your organization.
Required profile
Experience
Level of experience:Senior (5-10 years)
Spoken language(s):
English
Check out the description to know which languages are mandatory.