Senior Data Engineer

Work set-up: 
Full Remote
Contract: 
Experience: 
Senior (5-10 years)
Work from: 

Offer summary

Qualifications:

Bachelor's degree in Computer Science or related field., Over 5 years of experience in building secure, scalable cloud data infrastructure., Proficiency in Python, SQL, and data transformation libraries like pandas or PySpark., Experience with cloud deployment (preferably Azure) and data pipeline orchestration tools like Apache Airflow or Prefect..

Key responsibilities:

  • Design, build, and manage scalable data pipelines for AI applications.
  • Collaborate with cross-functional teams to understand and implement data solutions.
  • Develop and optimize cloud databases and data storage systems.
  • Ensure data quality, security, and governance across all workflows.

Board of Innovation logo
Board of Innovation SME https://www.boardofinnovation.com/
51 - 200 Employees
See all jobs

Job description

Senior Data Engineer

About Us

BOI is pioneering the field of AI strategy and applied AI, partnering with largescale, ambitious businesses around the world to ideate, design, and build AI products that transform the way they work. We don’t just help organizations imagine what’s possible we build the software that turns that vision into reality.

We work across different sectors and functions and have built a worldclass team that unites commercial expertise, smart strategy, and technology. If you’re passionate about building enterprise AI products from 0>1 and leading projects at the forefront of AI innovation, we’d love to meet you!

About the Role

We are hiring a Lead Data Engineer to design, build, and manage scalable data pipelines that support our AIpowered tools and applications, including agentic and predictive AI. You will integrate, harmonize, and transform data from disparate sources to ensure it is ready for use in foundational model integrations. This is a handson role where you will be responsible for building and implementing systems from the ground up. You would write productionlevel code while defining processes and best practices for future team growth.

You will

  • Work together with a collaborative international team of engineers, designers and solution leads to understand customer business needs and translate them into endtoend solutions.
  • Develop and manage ETL pipelines to extract, transform, and load data from various internal and external sources into harmonized datasets.
  • Design, optimize, and maintain cloud databases and data storage systems (PostgresSQL, MongoDB, Object storage).
  • Ensure data integrity, quality, and security across all pipelines and workflows, and enforce data access and governance systems.
  • Collaborate with AI Engineers to prepare data for use in model workflows (e.g. vector embeddings).
    • You have

      • 5+ years of experience building secure and scalable cloud data infrastructure, and a degree in Computer Science or related field.
      • Experience in Python, SQL, and data transformation libraries like pandas or PySpark.
      • Handson experience with data pipeline orchestration tools like Apache Airflow or Prefect.
      • Solid understanding of database design and optimization for relational and nonrelational databases.
      • Experience with deployment on one of the major cloud providers (preferably Azure).
      • Advanced English skills, both written and verbal, with the ability to communicate effectively in an international team.
        • You are

          • Experience working in a consulting or agency environment on projectbased work.
          • Passionate about working with data and building reliable scalable systems.
          • Familiar with DevOps practices in CICD for infraascode (e.g. terraform, helm, argo), container orchestration (e.g. docker, k8s), and deployment automation.
          • Passionate about continuously improving code quality by performing code reviews, and following best practices for testing and debugging.
          • Able to take charge of a project from conception till productionlevel customer deployment.
          • Willing to work with the team to develop knowledge around new technology as well as software development best practices.
            • Job evolution

              • Within your first month, you will integrate with the team, help set up best development practices and initial data pipelines.
                  • Within your third month, you will be actively involved in a client project, scoping the project and implementing user stories.
                  • Within your six months, you will be involved in multiple client projects, having already brought one from inception to a scalable client deployment. You will be actively working with the team to develop internal software primitives and IP.
                    • Note: We will start screening candidates after August 25th.

Required profile

Experience

Level of experience: Senior (5-10 years)
Spoken language(s):
English
Check out the description to know which languages are mandatory.

Other Skills

  • Teamwork
  • Communication
  • Problem Solving

Data Engineer Related jobs