DataMachine Learning Engineer Senior

Work set-up: 
Full Remote
Contract: 
Experience: 
Senior (5-10 years)

Offer summary

Qualifications:

8+ years of experience as a Data Engineer with at least 2 years in MLOps., Proficiency in Python for data transformation and large-scale processing., Deep understanding of vector databases, RAG architectures, and semantic retrieval workflows., Experience with cloud platforms like AWS or Azure for LLM deployments..

Key responsibilities:

  • Design and maintain scalable data pipelines for ingestion and transformation.
  • Develop workflows for extracting and retrieving semantic representations of data.
  • Create analytics and dashboards with natural language query capabilities.
  • Manage prompt engineering, model fine-tuning, and orchestration for conversational AI.

IDT Corporation logo
IDT Corporation Large http://www.idt.net
1001 - 5000 Employees
See all jobs

Job description

This is a fulltime work from home opportunity for a star DataML Engineer from LATAM.

IDT(www.idt.net) is an American telecommunications company founded in 1990 and headquartered in New Jersey. Today it is an industry leader in prepaid communication and payment services and one of the world’s largest international voice carriers. We are listed on the NYSE, employ over 1300 people across 20+ countries, and have revenues in excess of $1.5 billion.

We are looking for a skilled DataML Engineer to join our BI team and take an active role in designing, building, and maintaining the endtoend data pipeline, architecture and design that powers our warehouse, LLMdriven applications, and AIbased BI. If youre looking for a company that will give you the maximum flexibility in choosing a location to work, this opportunity is for you!

Responsibilities:
  • Design, develop, and maintain scalable data pipelines to support ingestion, transformation, and delivery into centralized feature stores, modeltraining workflows, and realtime inference services.
  • Build and optimize workflows for extracting, storing, and retrieving semantic representations of unstructured data to enable advanced search and retrieval patterns.
  • Architect and implement lightweight analytics and dashboarding solutions that deliver natural language query experience and AIbacked insights.
  • Define and execute processes for managing prompt engineering techniques, orchestration flows, and model finetuning routines to power conversational interfaces.
  • Oversee vector data stores and develop efficient indexing methodologies to support retrievalaugmented generation (RAG) workflows.
  • Partner with data stakeholders to gather requirements for languagemodel initiatives and translate into scalable solutions.
  • Create and maintain comprehensive documentation for all data processes, workflows and model deployment routines.
  • Should be willing to stay informed and learn emerging methodologies in data engineering, MLOps and LLM operations.


  • Requirements:
  • 8+ years of experience as a Data Engineer with 2+ years focused on MLOps.
  • Excellent English communication skills.
  • Effective oral and written communication skills with BI team and user community.
  • Demonstrated experience in utilizing python for data engineering tasks, including transformation, advanced data manipulation, and largescale data processing.
  • Deep understanding of vector databases and RAG architectures, and how they drive semantic retrieval workflows.
  • Skilled at integrating opensource LLM frameworks into data engineering workflows for endtoend model training, customization, and scalable inference.
  • Experience with cloud platforms like AWS or Azure Machine Learning for managed LLM deployments.
  • Handson experience with big data technologies including Apache Spark, Hadoop, and Kafka for distributed processing and realtime data ingestion.
  • Experience designing complex data pipelines extracting data from RDBMS, JSON, API and Flat file sources.
  • Demonstrated skills in SQL and PLSQL programming, with advanced mastery in Business Intelligence and data warehouse methodologies, along with handson experience in one or more relational database systems and cloudbased database services such as SnowflakeRedshift.
  • Understanding of software engineering principles and skills working on UnixLinuxWindows Operating systems, and experience with Agile methodologies.
  • Proficiency in version control systems, with experience in managing code repositories, branching, merging, and collaborating within a distributed development environment.
  • Interest in business operations and comprehensive understanding of how robust BI systems drive corporate profitability by enabling datadriven decisionmaking and strategic insights.


  • Pluses
  • Experience with vector databases such as DataStax AstraDB, and developing LLMpowered applications using popular open source frameworks like LangChain and LlamaIndex–including prompt engineering, retrievalaugmented generation (RAG), and orchestration of intelligent workflows.
  • Familiarity with evaluating and integrating opensource LLM frameworks–such as Hugging Face TransformersLLaMA4 across endtoend workflows, including finetuning and inference optimization.
  • Knowledge of MLOps tooling and CICD pipelines to manage model versioning and automated deployments.

  • Please attach CV in English.
    The interview process will be conducted in English.

    Only accepting applicants from LATAM.
  • Required profile

    Experience

    Level of experience: Senior (5-10 years)
    Spoken language(s):
    English
    Check out the description to know which languages are mandatory.

    Other Skills

    • Communication

    Related jobs