Match score not available

Intermediate Data Engineer - OP01505

extra holidays
Remote: 
Full Remote
Contract: 
Experience: 
Mid-level (2-5 years)
Work from: 

Offer summary

Qualifications:

4+ years of experience in data engineering, Advanced experience with unstructured files and legacy systems, Expertise in cloud storage solutions on Google Cloud Platform, Proficient in programming with Python for data manipulation, Familiarity with SQL and NoSQL databases.

Key responsabilities:

  • Manage data warehouse and pipeline processes
  • Migrate data to Google Cloud services
  • Design and maintain ETL/ELT pipelines
  • Collaborate on data mastering tool improvements
  • Evaluate and recommend technology stack improvements
Dev.Pro logo
Dev.Pro SME https://dev.pro/
501 - 1000 Employees
See more Dev.Pro offers

Job description

🟒 Join Dev.Pro - a US-based outsourcing software development company that has been delivering exceptional software experience to clients since 2011.

We invite a talented, organized, and reliable Data Engineer to join our team and play an important role in building a project that connects the art market with the digital-first financial ecosystem. Work with a global leader in artwork tokenization and blockchain-secured provenance, using your skills to contribute to innovative projects at the crossroads of the blockchain and art industries.

🟩 What's in it for you:

  • Engage in an exciting blockchain-related project transforming the art market
  • Work within a unique business domain that collaborates with world-class brands
  • Join a dynamic no-legacy startup and contribute to building processes, tools, and products from scratch

βœ… Is that you?

  • 4+ years of experience in data engineering, encompassing data extraction, transformation, and migration.
  • Advanced experience with data extraction from unstructured files and legacy systems.
  • Proven expertise in migrating data from file-based storage systems to cloud storage solutions, ideally on Google Cloud Platform.
  • Proficiency with relational databases, specifically MariaDB or MySQL, as well as cloud-native solutions such as Google Cloud Storage, Google BigQuery, and optionally Snowflake or Amazon Redshift.
  • Strong programming skills in Python, with a focus on data manipulation, automation, and re-implementing custom tools.
  • Extensive experience with ETL/ELT pipeline development and workflow orchestration tools (e.g., Apache Airflow, Luigi, Google Dataflow, Prefect).
  • Hands-on experience with batch processing frameworks and real-time data processing frameworks.
  • Experience with data pipeline development using programming languages, including batch processing implementation.
  • In-depth understanding of data modeling, data warehousing, and best practices for designing scalable data architectures.
  • Practical experience in developing or re-engineering data mastering tools for the purpose of data cleaning, standardization, and preparation.
  • Expertise in RDBMS functionalities, such as stored procedures, triggers, partitioning, indexes, and structural changes.
  • Ability to handle Personally Identifiable Information (PII) data within pipelines and data storage systems.
  • Experience with NoSQL databases, such as MongoDB, Cassandra, or HBase.
  • Experience with monitoring tools such as Prometheus, Grafana, and CloudWatch to oversee data pipelines and systems.
  • Knowledge of best practices in database management, performance optimization, data security, and ensuring consistency across distributed systems.
  • Ability to critically evaluate data architecture and provide strategic recommendations for infrastructure improvements.
  • Upper-Intermediate+ English level

Desirable:

  • Familiarity with JavaScript for maintaining or enhancing legacy systems and cross-functional integration.
  • Experience with ElasticSearch for indexing and querying large datasets.
  • Proficiency with analytical tools such as Tableau, Power BI, Looker, or similar platforms for data visualization and insights generation.
  • Interest or background in the art industry, particularly related to digital asset management and tokenization.
  • Demonstrated ability to collaborate in cross-functional teams and contribute to multidisciplinary projects.
  • Experience with PostgreSQL and understanding its application in data engineering environments.
  • Knowledge of specific services related to data engineering, including key metrics and business processes relevant to the industry domain.
  • Experience with MLOps tools and practices to streamline machine learning deployment and operations.
  • Basic understanding of existing machine learning models and algorithms.

🧩Key responsibilities and your contribution

As a Data Engineer, you will contribute significantly to the data team's efforts on a rapidly expanding project, helping to manage increasing workloads while maintaining high standards of quality and efficiency. Your key responsibilities will include:

  • Taking full responsibility for the data warehouse and pipeline, including planning, coding, reviews, and delivery to the production environment.
  • Migrating data from existing file storage systems to the Google Cloud Platform, including Google Cloud Storage and BigQuery.
  • Designing, developing, and maintaining ETL/ELT pipelines to support data migration and integration.
  • Collaborating with team members to re-implement existing custom data mastering tools, with a focus on improving data cleaning and standardization capabilities.
  • Conducting thorough evaluations of the existing technology stack and providing data-driven recommendations for improvements, including re-evaluating database solutions and orchestration tools.
  • Developing a new scraper system to extract and aggregate data from diverse external sources, ensuring integration with existing platforms.
  • Ensuring the integrity, consistency, and quality of data through optimized processes and validation protocols.
  • Working closely with a small, dynamic team to ensure that project milestones are met effectively, with an emphasis on scalability, reliability, and sustainability of solutions.

🎾 What's working at Dev.Pro like?

At Dev.Pro, we appreciate your contribution and the value you add to projects. We are a company with an ambitious and creative mindset whose corporate culture is based on respect for diversity and uniqueness, as well as enablement of equal opportunities for all. By joining Dev.Pro, you'll feel what it's like to grow with professionals who support your journey.

βœ… Get 30 paid rest days per year to use as holidays/vacation/other on the desired and requested dates
βœ… 5 sick leave days, up to 60 days of medical leave, and up to 6 days of leave per year due to family reasons (i.e., wedding/funeral/baby birth)
βœ… Get a health insurance package fully compensated by Dev.Pro
βœ… Join fun online activities and team-building events
βœ… Get continuous remote HR, payroll support, and overtime coverage
βœ… Join English/Polish lessons
βœ… Grow your expertise with mentorship support and DP University

Our next steps:

βœ… Submit a CV in English β€” βœ… Intro call with a Recruiter β€” βœ… Internal interview β€” βœ… Client interview β€” βœ… Offer

Interested? Find out more:

πŸ“‹How we work

πŸ’» LinkedIn Page

πŸ“ˆ Our website

πŸ’»IG Page

Required profile

Experience

Level of experience: Mid-level (2-5 years)
Spoken language(s):
EnglishEnglish
Check out the description to know which languages are mandatory.

Other Skills

  • Collaboration
  • Communication
  • Analytical Thinking
  • Problem Solving

Data Engineer Related jobs