Artwork

Content provided by Roger Basler de Roca. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Roger Basler de Roca or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://ppacc.player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!

Can you believe your AI? Detecting Hallucinations in Language Models

6:06
 
Share
 

Manage episode 446444128 series 3153807
Content provided by Roger Basler de Roca. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Roger Basler de Roca or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://ppacc.player.fm/legal.

In this episode, we delve into the intriguing challenge of "hallucinations" in large language models (LLMs)—responses that are grammatically correct but factually incorrect or nonsensical. Drawing from a groundbreaking paper, we explore the concept of epistemic uncertainty, which stems from a model's limited knowledge base.

Unlike previous approaches that often only measure the overall uncertainty of a response, the authors introduce a new metric that distinguishes between epistemic and aleatoric (random) uncertainties. This distinction is crucial for questions with multiple valid answers, where high overall uncertainty doesn't necessarily indicate a hallucination.

Experimentally, the authors demonstrate that their method outperforms existing approaches, especially in datasets that include both single-answer and multiple-answer questions. Their method is particularly effective in high-entropy questions, where the model is uncertain about the correct answer.

Join us as we unpack this promising approach to detecting hallucinations in LLMs, grounded in solid theoretical foundations and proven effective in practice.

This episode is based on the paper: Yasin Abbasi-Yadkori, Ilja Kuzborskij, András György, Csaba Szepesvári. "To Believe or Not to Believe Your LLM", ArXiv:2406.02543v1, 2024, it can be found here.

Disclaimer: This podcast is generated by Roger Basler de Roca (contact) by the use of AI. The voices are artificially generated and the discussion is based on public research data. I do not claim any ownership of the presented material as it is for education purpose only.

  continue reading

50 episodes

Artwork
iconShare
 
Manage episode 446444128 series 3153807
Content provided by Roger Basler de Roca. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Roger Basler de Roca or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://ppacc.player.fm/legal.

In this episode, we delve into the intriguing challenge of "hallucinations" in large language models (LLMs)—responses that are grammatically correct but factually incorrect or nonsensical. Drawing from a groundbreaking paper, we explore the concept of epistemic uncertainty, which stems from a model's limited knowledge base.

Unlike previous approaches that often only measure the overall uncertainty of a response, the authors introduce a new metric that distinguishes between epistemic and aleatoric (random) uncertainties. This distinction is crucial for questions with multiple valid answers, where high overall uncertainty doesn't necessarily indicate a hallucination.

Experimentally, the authors demonstrate that their method outperforms existing approaches, especially in datasets that include both single-answer and multiple-answer questions. Their method is particularly effective in high-entropy questions, where the model is uncertain about the correct answer.

Join us as we unpack this promising approach to detecting hallucinations in LLMs, grounded in solid theoretical foundations and proven effective in practice.

This episode is based on the paper: Yasin Abbasi-Yadkori, Ilja Kuzborskij, András György, Csaba Szepesvári. "To Believe or Not to Believe Your LLM", ArXiv:2406.02543v1, 2024, it can be found here.

Disclaimer: This podcast is generated by Roger Basler de Roca (contact) by the use of AI. The voices are artificially generated and the discussion is based on public research data. I do not claim any ownership of the presented material as it is for education purpose only.

  continue reading

50 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Quick Reference Guide

Listen to this show while you explore
Play