Logo for NTT DATA

Azure Databricks and Oracle Tester

Roles & Responsibilities

  • 5+ years of hands-on experience in data testing, QA, or data quality engineering in data warehouse, lake, or analytics modernization projects.
  • Proven experience validating data pipelines involving Oracle (source/warehouse) and Databricks (target lakehouse), including source-to-target reconciliations.
  • Proficiency in Oracle SQL and Python/PySpark for automated data quality checks; familiarity with Delta Lake concepts.
  • Experience with Agile/Scrum, CI/CD integrations (e.g., Azure DevOps, Jira/Git), and strong ETL/ELT and data warehousing fundamentals.

Requirements:

  • Design and execute data test strategies, plans, and test cases for ingestion, transformation, and curated layers; validate data at rest and in motion across Oracle and Databricks.
  • Perform comprehensive source-to-target reconciliation (record counts, hashes, aggregates, sampling) and validate transformation logic, incremental loads, and CDC/recovery scenarios.
  • Build and maintain automated data quality checks using SQL and Python (PySpark); develop reusable validation utilities and integrate tests into CI/CD pipelines (Azure DevOps, GitHub, Jenkins).
  • Collaborate with data engineers and business stakeholders to troubleshoot pipelines, document test evidence, governance artifacts, and support issue resolution.

Job description



NTT DATA strives to hire exceptional, innovative and passionate individuals who want to grow with us. If you want to be part of an inclusive, adaptable, and forward-thinking organization, apply now.

We are currently seeking a Azure Databricks and Oracle Tester to join our team in Remote, Texas (US-TX), United States (US).

Role Summary (Role Description)
We are seeking a Data Tester / Data Quality Engineer to support a data modernization initiative, validating data pipelines and reconciling legacy (Oracle) and modern (Databricks/Lakehouse) environments. This role will be responsible for designing and executing data validation tests, ensuring data completeness, accuracy, consistency, and timeliness as datasets migrate and pipelines are rebuilt or replatformed.
You'll work in an Agile delivery model with cross-functional teams to validate ETL/ELT transformations, ensure correct business rule implementation, and build repeatable, automated validation frameworks that scale across domains.
Key Responsibilities
  • Data Testing & Validation
  • Design and execute data test strategies, test plans, and test cases for ingestion, transformation, and curated layers.
  • Validate data at rest and in motion across Oracle source systems and Databricks target platforms.
  • Perform source-to-target reconciliation, including:
  • record counts, checksum/hashing, aggregates, sampling
  • null/constraint checks, referential integrity, duplicates
  • transformation logic validation (business rules, SCD logic, dedup, enrichment)
  • Validate incremental loads, CDC patterns, and rerun/recovery scenarios.
  • Automation & Frameworks
  • Build and maintain automated data quality checks using SQL and/or Python (e.g., PySpark).
  • Develop reusable data validation utilities and parameterized scripts to reduce manual effort.
  • Integrate data tests into CI/CD pipelines where applicable (e.g., Azure DevOps, GitHub, Jenkins).
  • Defect Management & Collaboration
  • Log, triage, and manage defects with clear reproduction steps and root-cause hints.
  • Collaborate with data engineers to troubleshoot pipeline failures and data anomalies.
  • Partner with business stakeholders/analysts to confirm expected outcomes and acceptance criteria.
  • Documentation & Governance Support
  • Document test evidence, reconciliation results, and sign-off artifacts for releases.
  • Support data governance objectives (quality KPIs, issue tracking, lineage/metadata readiness).

Required Skills & Qualifications Experience:
  • 5 or more years of hands-on experience in data testing / QA / data quality engineering in data warehouse, lake, or analytics modernization initiatives.
  • Demonstrated experience validating data pipelines involving Oracle (source or warehouse) and Databricks (target lakehouse).
Technical Skills
  • Oracle SQL: complex queries, joins, aggregations, performance-aware validation queries.
Databricks:
  • Experienced with of Spark/Databricks SQL
  • Familiarity with Delta Lake concepts (tables, merges/upserts, partitions, time travel helpful)
  • Data validation techniques:
  • reconciliations, profiling, anomaly detection basics
  • test data creation and boundary testing for transformations
  • Python (preferred) and/or PySpark for automation and scalable validations.
  • Strong understanding of ETL/ELT concepts, data warehousing fundamentals, and dimensional modeling basics.
  • Tools & Ways of Working
  • Familiarity with Agile/Scrum, user stories, acceptance criteria, and sprint-based testing.
  • Experience with defect tracking tools (e.g., Jira, Azure DevOps).
  • Version control basics (e.g., Git).

Preferred / Nice-to-Have Skills
  • Data quality frameworks or tooling (e.g., Great Expectations, Deequ, Soda, dbt tests).
  • Experience validating orchestration outputs (e.g., ADF, Airflow, Informatica, ODI).
  • Experience with cloud platforms (Azure/AWS/GCP) and storage layers (ADLS/S3/GCS).
  • Understanding of CDC patterns and validation (GoldenGate, Debe)


About NTT DATA:

NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com


NTT DATA endeavors to make https://us.nttdata.com accessible to any and all users. If you would like to contact us regarding the accessibility of our website or need assistance completing the application process, please contact us at https://us.nttdata.com/en/contact-us. This contact information is for accommodation requests only and cannot be used to inquire about the status of applications. NTT DATA is an equal opportunity employer. Qualified applicants will receive consideration for employment without regard to race, color, religion, sex, sexual orientation, gender identity, national origin, disability or protected veteran status. For our EEO Policy Statement, please click here. If you'd like more information on your EEO rights under the law, please click here. For Pay Transparency information, please click here.

Related jobs

Other jobs at NTT DATA

We help you get seen. Not ignored.

We help you get seen faster β€” by the right people.

πŸš€

Auto-Apply

We apply for you β€” automatically and instantly.

Save time, skip forms, and stay on top of every opportunity. Because you can't get seen if you're not in the race.

✨

AI Match Feedback

Know your real match before you apply.

Get a detailed AI assessment of your profile against each job posting. Because getting seen starts with passing the filters.

Upgrade to Premium. Apply smarter and get noticed.

Upgrade to Premium

Join thousands of professionals who got noticed and hired faster.