Logo for Worth AI

Data Platform Engineer

Roles & Responsibilities

  • Expertise with graph databases (Neo4j, AWS Neptune, or TigerGraph) and query languages Cypher or Gremlin
  • Experience with entity resolution or record linkage (e.g., Senzing, Quantexa, or probabilistic matching models)
  • Ability to design flexible ontologies for evolving regulatory data (PEP, sanctions, AML/KYC data structures)
  • Experience building centralized data platforms at scale with API first design (GraphQL/REST) optimized for graph traversals

Requirements:

  • Architect and implement entity resolution logic to deduplicate and link disparate data points into unified Golden Records and maintain a global knowledge graph and ontology for ownership and risk relationships
  • Design and maintain a hybrid storage strategy bridging graph databases with document and search stores to support rich metadata and risk content
  • Develop scalable data services and batch/streaming pipelines using AWS native tooling to ingest, transform, and serve data across the company
  • Own the reliability and performance of the API first data platform, including monitoring, on call, data quality, lineage, governance, and CI/CD automation

Job description

Worth AI, a leader in the computer software industry, is looking for a talented and experienced Data Platform Engineer to join their innovative team. At Worth AI, we are on a mission to revolutionize decision-making with the power of artificial intelligence while fostering an environment of collaboration, and adaptability, aiming to make a meaningful impact in the tech landscape.. Our team values include extreme ownership, one team and creating reaving fans both for our employees and customers.

As a Data Platform Engineer, you will design, build, and operate the core data services that power our products and analytics. You’ll own end-to-end data pipelines and API services that ingest, process, and expose high-quality data to internal customers (data science, analytics, product, and other engineering teams) and external partners.

You’ll be part of a small, high-impact team that treats the data platform as a product with strong SLAs, and reliable self-service for internal and external users.

Responsibilities

What you’ll do:

  • Architect and implement entity resolution logic to de-duplicate and link disparate data points into unified "Golden Records" for businesses and individuals
  • Design and maintain a high-performance global business knowledge graph and ontology to map complex ownership chains, UBOs, and hidden risk relationships across international borders
  • Implement a hybrid storage strategy that bridges graph databases for relationship mapping with document and search stores for rich metadata and adverse media content
  • Optimize the platform for real-time risk assessment, ensuring the ability to traverse multiple levels of ownership in milliseconds to support automated "Go/No-Go" onboarding decisions
  • Design and build scalable data services and APIs for ingesting, transforming, and serving data across the company
  • Develop and maintain batch and streaming data pipelines using modern data processing frameworks and AWS cloud-native tooling
  • Own the reliability, performance, and API first data platform, including monitoring, alerting, and on-call where appropriate
  • Implement best practices for data modeling, quality, lineage, and governance to ensure trustworthy, well-documented datasets
  • Work closely with data scientists, analysts, and application engineers to understand their needs and translate them into robust platform capabilities
  • Drive automation and standardization through CI/CD, model as a service, and reproducible environments
  • Help define and evolve the architecture of our data platform as a true internal service with clear contracts, SLAs, and versioned APIs

Requirements

    • Expertise in Graph Ecosystems: Hands-on experience with Graph databases (e.g., Neo4j, AWS Neptune, or TigerGraph) and query languages like Cypher or Gremlin
    • Identity & Linkage Mastery: Proven experience with Entity Resolution or Record Linkage (e.g., using tools like Senzing, Quantexa, or custom probabilistic matching models)
    • Schema Design: Ability to design flexible ontologies that handle evolving regulatory data (e.g., changing PEP definitions or Sanction list formats)
    • API Performance for Graphs: Experience building GraphQL or REST APIs specifically optimized for graph traversals and deep-tree lookups
    • Experience building centralized data platforms or “data-as-a-service” offerings at scale (e.g., at a large tech or cloud-native company)
    • Strong software engineering skills in at least one language commonly used for data and services (e.g., Python, Java, Go, Rust)
    • Hands-on experience building data pipelines and ETL/ELT workflows on a major cloud provider (AWS preferred)
    • Experience with modern data stack tools such as Spark/Flink, Kafka/Kinesis, Airflow/managed schedulers, and data warehouses (e.g., Snowflake, Redshift, BigQuery, Databricks)
    • Familiarity with DevOps practices: CI/CD, containerization (Docker), orchestration (Kubernetes), and infrastructure-as-code (Terraform)
    • Strong focus on observability (metrics, logs, traces), resilience, and building early warning signals
    • Comfort collaborating cross-functionally and communicating clearly with both technical and non-technical stakeholders.

Nice to Have

    • Background supporting machine learning or real-time decisioning use cases from a platform point of view
    • Compliance Domain Knowledge: Understanding of AML, CTF, and KYC/KYB data structures (e.g., LEIs, ISO 20022)
    • Geospatial Data: Experience handling global address normalization and geospatial indexing for risk detection

** All Remote Hires - will be required to travel to Orlando, Florida at least twice per year for Town Halls and team collaboration in addition to orientation in Orlando, Florida

Benefits

    • Health Care Plan (Medical, Dental & Vision)
    • Retirement Plan (401k, IRA)
    • Life Insurance
    • Flexible Paid Time Off
    • 9 paid Holidays
    • Family Leave
    • Work From Home
    • Free Food & Snacks (Orlando)
    • Wellness Resources

Data Engineer Related jobs

Other jobs at Worth AI

We help you get seen. Not ignored.

We help you get seen faster — by the right people.

🚀

Auto-Apply

We apply for you — automatically and instantly.

Save time, skip forms, and stay on top of every opportunity. Because you can't get seen if you're not in the race.

AI Match Feedback

Know your real match before you apply.

Get a detailed AI assessment of your profile against each job posting. Because getting seen starts with passing the filters.

Upgrade to Premium. Apply smarter and get noticed.

Upgrade to Premium

Join thousands of professionals who got noticed and hired faster.