Using Custom Evaluators During Training | Datawizz

Datawizz · Beginner ·🧠 Large Language Models ·4mo ago
This video demonstrates how to implement custom evaluation metrics during model training in Datawizz, allowing you to monitor domain-specific performance metrics alongside standard training/validation loss. Technical Overview: In this example, I train a conversation summarization model (Qwen-0.6B base) and track ROUGE scores during training. The feature supports both built-in metrics and custom evaluation functions, including LLM-as-judge implementations. Key Features Covered: - Configuring custom evaluators in the training pipeline - Real-time metric visualization during training epochs - A…
Watch on YouTube ↗ (saves to browser)

Chapters (6)

Feature introduction
0:34 Dataset structure
0:52 Training configuration
1:15 Custom evaluator setup
2:03 Metric analysis during training
2:40 Evaluation sample inspection
5 Levels of AI Agents - From Simple LLM Calls to Multi-Agent Systems
Next Up
5 Levels of AI Agents - From Simple LLM Calls to Multi-Agent Systems
Dave Ebbelaar (LLM Eng)