Listen "Measuring LLMs with Jodie Burchell"
Episode Synopsis
How do you measure the quality of a large language model? Carl and Richard talk to Dr. Jodie Burchell about her work measuring large language models for accuracy, reliability, and consistency. Jodie talks about the variety of benchmarks that exist for LLMs and the problems they have. A broader conversation about quality digs into the idea that LLMs should be targeted to the particular topic area they are being used for - often, smaller is better! Building a good test suite for your LLM is challenging but can increase your confidence that the tool will work as expected.
More episodes of the podcast .NET Rocks!
Energy Geek Out 2025
01/01/2026
Space Geek Out 2025
25/12/2025
The Role of AI in Software Development
18/12/2025
Building an AI App with Calum Simpson
04/12/2025
More Sustainable Software with Tom Kerkhove
27/11/2025
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.