Logo for Xsolla

Data Scienist

Roles & Responsibilities

  • 5+ years of experience in data science, with 3+ years in Spark framework
  • Strong SQL and Python skills, with proven experience building ETL/ELT at scale
  • Experience with data pipeline orchestration (Airflow, Prefect, dbt, or similar)
  • Solid understanding of data modeling (Kimball, Data Vault, or hybrid)

Requirements:

  • Design, build, and optimize data pipelines and ETL workflows in Snowflake (Snowpark, Streams/Tasks, Snowpipe); develop scalable data models, user 360 views, churn prediction, and recommendation engine inputs; lead integration across data sources (MySQL, BigQuery, Redis, Kafka, GCP Storage, API Gateway); implement CI/CD and data quality checks
  • Mentor junior data engineers; partner with Data Science, ML, and Backend teams to productionize ML features in Snowflake; ensure data governance, privacy, and PII compliance; collaborate with stakeholders to translate requirements
  • Tune algorithm performance; establish data partitioning, clustering, and materialized views; build dashboards and monitors for pipeline health, job success, and data latency metrics
  • Establish naming conventions, data lineage, and metadata standards; lead code reviews and documentation; contribute to evolving data mesh and streaming architecture vision

Job description

ABOUT US

Xsolla is a global commerce company with robust tools and services to help developers solve the inherent challenges of the video game industry. From indie to AAA, companies partner with Xsolla to help them fund, distribute, market, and monetize their games. Grounded in the belief in the future of video games, Xsolla is resolute in the mission to bring opportunities together, and continually make new resources available to creators. Headquartered and incorporated in Los Angeles, California, Xsolla operates as the merchant of record and has helped over 1,500+ game developers to reach more players and grow their businesses around the world. With more paths to profits and ways to win, developers have all the things needed to enjoy the game.

For more information, visit xsolla.com.

Responsibilities
  • 1. Architecture & Development
  • Design, build, and optimize data pipelines and ETL workflows in Snowflake using Snowpark, Streams/Tasks, and Snowpipe.
  • Develop scalable data models, Algorithm supporting user 360 views, churn prediction, and recommendation engine inputs.
  • Lead integration across data sources: MySQL, BigQuery, Redis, Kafka, GCP Storage, and API Gateway.
  • Implement CI/CD for data pipelines using Git, dbt, and automated testing.
  • Define data quality checks and auditing pipelines for ingestion and transformation layers.
  • 2. Leadership & Collaboration
  • Mentor and guide junior data engineers on data modeling, performance tuning, and Snowflake best practices.
  • Partner with Data Science, ML, and Backend teams to productionize machine learning features in Snowflake.
  • Work closely with Legal, Security, and Infrastructure teams to ensure compliance, privacy, and governance of user data (PII).
  • Collaborate with the Director of Data Platforms and product stakeholders to translate business requirements into technical specifications.
  • 3. Performance & Scalability
  • Tune algorithm performance.
  • Establish data partitioning, clustering, and materialized views for fast query execution.
  • Build dashboards and monitors for pipeline health, job success, and data latency metrics (e.g., via Looker, Tableau, or Snowsight).
  • 4. Governance & Best Practices
  • Establish and enforce naming conventions, data lineage, and metadata standards across schemas.
  • Lead code reviews, enforce documentation standards, and manage schema versioning.
  • Contribute to the company’s evolving data mesh and streaming architecture vision.

  • Qualification & Skills
  • 5+ years of experience in Data Scientist, with 3+ years in Spark framework.
  • Strong SQL and Python skills, with proven experience building ETL/ELT at scale.
  • Deep understanding of algorithm performance tuningquery optimization, and warehouse orchestration.
  • Experience with data pipeline orchestration (Airflow, Prefect, dbt, or similar).
  • Solid understanding of data modeling (Kimball, Data Vault, or hybrid).
  • Proficiency in KafkaGCP, or AWS for real-time or batch ingestion.
  • Familiarity with API-based data integration and microservice architectures.
  • Preferred
  • Experience lead machine learning teams or/and deploying ML feature pipelines.
  • Background in ad-tech, gaming, or e-commerce recommendation systems.
  • Familiarity with data contracts and feature stores (Feast, Tecton, or custom-built).
  • Experience managing small data engineering teams and setting technical direction
  • Strong ownership and ability to work autonomously in a fast-paced environment.
  • Excellent cross-functional communication — can translate between engineering and business.
  • Hands-on problem solver who balances velocity with reliability.
  • Collaborative mentor who raises the bar for team quality and discipline
  • Related jobs

    Other jobs at Xsolla

    We help you get seen. Not ignored.

    We help you get seen faster — by the right people.

    🚀

    Auto-Apply

    We apply for you — automatically and instantly.

    Save time, skip forms, and stay on top of every opportunity. Because you can't get seen if you're not in the race.

    AI Match Feedback

    Know your real match before you apply.

    Get a detailed AI assessment of your profile against each job posting. Because getting seen starts with passing the filters.

    Upgrade to Premium. Apply smarter and get noticed.

    Upgrade to Premium

    Join thousands of professionals who got noticed and hired faster.