Minimum 6 years of experience as a Data Engineer with large-scale data architectures., Strong expertise in SQL and advanced proficiency in Python for ETL and data automation., In-depth experience with Apache Spark and Kafka for big data processing and real-time data streaming., BSc in Computer Science, Data Engineering, or a relevant field..
Key responsibilities:
Design, build, and manage ETL/ELT pipelines using tools like AWS Glue and Azure Data Factory.
Develop scalable real-time and batch data processing solutions using Apache Spark and Kafka.
Collaborate with data analysts and engineers to define data models and ensure data integrity throughout the pipeline.
Monitor and troubleshoot pipeline performance and automate workflows in cloud-native environments.
Report This Job
Help us maintain the quality of our job listings. If you find any issues with this job post, please let us know.
Select the reason you're reporting this job:
At SPACE44, we specialize in empowering tech startups and established technology companies in the US and Europe by providing high-level engineers and tailored HR solutions. Our services are designed to cater to a variety of needs, from creating stable hiring pipelines and efficient onboarding processes to performance management and comprehensive administrative support—all at no extra charge.
We utilize proven methodologies and Developer Playbooks to enhance team cohesion and productivity, ensuring seamless integration of remote and hybrid work settings. Our scalable and flexible approach adapts team sizes to meet diverse project demands and organizational growth, making us a perfect partner for startups and large enterprises alike.
Our deep understanding of remote work environments allows us to unlock significant potential and flexibility for businesses. We address the challenges of high costs and dependency on traditional suppliers by offering alternative engagement models that enhance cost efficiency and reduce reliance on obsolete practices. We prioritize creating high-performing, cohesive teams that drive innovation and success, even in complex, multi-layered organizations.
By breaking down traditional barriers, we foster an environment where expertise can thrive and innovation can flourish across various scales of business operations. At SPACE44, we are dedicated to engineering ecosystems where your team can excel, ensuring that your vision becomes reality. Join us in paving the path to the future of work, where empathy and technical prowess go hand in hand, enhancing security and integrating seamlessly with your existing structures.
At SPACE44, we empower tech startups and innovative companies across the US and Europe by delivering top-tier software engineers and comprehensive HR solutions. Unlike traditional staffing agencies, we accelerate innovation by providing performant hiring pipelines, efficient onboarding processes, performance management, and administrative assistance as an end-to-end service. Our unique approach ensures precise integration and high team performance in full-remote, nearshore, and offshore settings.
We enhance your work-life balance by offering opportunities that align your personal and professional aspirations. Work remotely to seamlessly integrate your career with your lifestyle and bring everything into a flow. We offer you a dynamic and supportive environment where you can develop and grow. Join a forward-thinking team that values the joy of work, openness, and cross-cultural collaboration. At SPACE44, let's shape the future of modern work together.
Who We Are Looking For?
We’re building our talent pool of outstanding Data Engineers for upcoming client projects. This is not an active position tied to a current project, but a proactive opportunity to become part of our expert network at SPACE44. When a suitable project arises, you'll be among the first we reach out to.
We’re always on the lookout for professionals who bring deep expertise, remote collaboration experience, and a passion for shaping the future of software and product development. If you enjoy solving complex problems and want to be considered for impactful future roles, we’d love to connect.
Key Responsibilities:
Design, build, and manage ETL/ELT pipelines using tools like AWS Glue, Azure Data Factory, and Apache Airflow
Develop scalable real-time and batch data processing solutions using Apache Spark and Kafka
Write optimized, production-grade SQL queries and perform performance tuning across data systems
Build and maintain data lakes and warehouses, including platforms like Snowflake and BigQuery
Work with structured and semi-structured data across relational and NoSQL databases (e.g., MongoDB, Cassandra)
Collaborate with data analysts, engineers, and product teams to define data models and architecture
Ensure data integrity, quality, and lineage throughout the pipeline
Automate workflows, testing, and deployment for data systems in cloud-native environments
Monitor and troubleshoot pipeline performance and reliability
Requirements:
Minimum 6 years of experience as a Data Engineer working with large-scale data architectures
Strong expertise in SQL for data manipulation, optimization, and analytics
Advanced proficiency in Python for ETL, scripting, and data automation
In-depth experience with Apache Spark for big data processing and distributed computing
Proven experience with Kafka for real-time data streaming and event processing
Hands-on experience with cloud-native ETL tools such as AWS Glue and Azure Data Factory
Familiarity with data warehousing platforms like Snowflake or BigQuery
Knowledge of NoSQL databases like MongoDB or Cassandra
Comfortable working in modern cloud environments and Agile remote teams
BSc in Computer Science, Data Engineering, or a relevant field
Nice to Have:
Experience with Apache Airflow or similar orchestration frameworks
Familiarity with GCP tools like BigQuery or Dataflow
Exposure to legacy big data systems such as Hadoop in hybrid environments
Understanding of data governance, security, and compliance in cloud data workflows
Knowledge of data modeling techniques for both OLAP and OLTP use cases
Interest in emerging technologies in real-time analytics and data mesh architecture
Embark on your SPACE44 journey with a few simple steps:
Hireflix Chat (20 min): Share how you work and what matters most to you
HR Sync (45 min): Connect your experience with our goals
Psychometric Insight: Take a 1-hour assessment exploring your strengths and working style
Benefits
Flexible Work Schedule: Enjoy a non-linear workday designed to enhance productivity and maintain a harmonious work-life balance, with core hours for team collaboration.
Professional Growth: Access advanced training opportunities in data science and machine learning to boost your career prospects.
Innovative Projects and Tools: Engage in cutting-edge projects using the latest tools and technologies within a progressive remote work environment.
Competitive Income: Receive a competitive income with regular performance reviews and potential raises every six months.
Global Team Dynamics: Collaborate with a diverse, international team that values openness and teamwork.
Required profile
Experience
Spoken language(s):
English
Check out the description to know which languages are mandatory.