Job Description:
• Design and execute test plans and test cases for AI/LLM-based applications
• Perform functional, regression, and performance testing for AI models and LLM integrations
• Validate prompt engineering techniques to ensure accurate and relevant model outputs
• Identify, track, and report defects in AI behavior and LLM responses
• Collaborate with AI/ML engineers to analyze model limitations and improve quality
• Develop automated testing frameworks for AI workflows and output validation
• Monitor model performance metrics such as accuracy, relevance, bias, and reproducibility
• Stay updated on the latest trends in AI/LLM testing and best practices
Required Skills:
• 5–6 years of experience in software testing, including 2+ years in AI/ML or LLM testing
• Strong understanding of Large Language Models (LLMs) and evaluation methods
• Hands-on experience in prompt engineering and generative AI output validation
• Familiarity with AI testing frameworks and model validation techniques
• Proficiency in Python, Java, or similar scripting languages
• Experience with automation tools like Selenium or PyTest is a plus
• Strong analytical, problem-solving, and documentation skills
Preferred Qualifications:
• BE / BTech / MCA
• Experience with AI models, NLP applications, or generative AI platforms
• Knowledge of AI ethics and responsible AI testing practices