Loading video player...
In this episode of XgridTalks, Ahmad Anas (Senior Technical Writer) speaks with M. Abu Bakkar Bin Akmal (DevOps & AI Systems Engineer) at Xgrid to break down one of the most overlooked parts of building AI products — continuous testing and evaluation. They walk through: • Why LLM quality is a continuous cycle — not a one-time check • How to test early using golden datasets and human review • Running full end-to-end and adversarial tests before launch • Using A/B testing and behavioral monitoring in production • Constant re-evaluation as users and requirements evolve If you’re deploying AI in real business environments, this testing mindset is critical. #LLM #GenAI #AITesting #MachineLearning #AIEngineering #MLOps #DigitalTransformation #EnterpriseAI #XgridTalks #ArtificialIntelligence