Go offline with the Player FM app!
Podcasts Worth a Listen
SPONSORED


1 Encore: Will Poulter, Dave Beran, and The Bear 52:22
[QA] OLMOTRACE: Tracing Language Model Outputs Back to Trillions of Training Tokens
Manage episode 476158571 series 3524393
OLMOTRACE is a real-time system that traces language model outputs to their training data, enabling users to explore fact-checking, hallucination, and creativity in language models.
https://arxiv.org/abs//2504.07096
YouTube: https://www.youtube.com/@ArxivPapers
TikTok: https://www.tiktok.com/@arxiv_papers
Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016
Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
2341 episodes
Manage episode 476158571 series 3524393
OLMOTRACE is a real-time system that traces language model outputs to their training data, enabling users to explore fact-checking, hallucination, and creativity in language models.
https://arxiv.org/abs//2504.07096
YouTube: https://www.youtube.com/@ArxivPapers
TikTok: https://www.tiktok.com/@arxiv_papers
Apple Podcasts: https://podcasts.apple.com/us/podcast/arxiv-papers/id1692476016
Spotify: https://podcasters.spotify.com/pod/show/arxiv-papers
2341 episodes
All episodes
×
1 [QA] Sampling from Your Language Model One Byte at a Time 7:05

1 Sampling from Your Language Model One Byte at a Time 13:35

1 [QA] Don't throw the baby out with the bathwater: How and why deep learning for ARC 7:44

1 Don't throw the baby out with the bathwater: How and why deep learning for ARC 32:30

1 [QA] What Happens During the Loss Plateau? Understanding Abrupt Learning in Transformers 7:18

1 What Happens During the Loss Plateau? Understanding Abrupt Learning in Transformers 19:43

1 [QA] MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning Attention 8:28

1 MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning Attention 25:05

1 [QA] Aligned Novel View Image and Geometry Synthesis via Cross-modal Attention Instillation 8:10

1 Aligned Novel View Image and Geometry Synthesis via Cross-modal Attention Instillation 16:59

1 [QA] TreeRL: LLM Reinforcement Learning with On-Policy Tree Search 7:17

1 TreeRL: LLM Reinforcement Learning with On-Policy Tree Search 19:00

1 [QA] Solving Inequality Proofs with Large Language Models 8:20

1 Solving Inequality Proofs with Large Language Models 23:49

1 [QA] Reinforcement Learning Teachers of Test Time Scaling 7:54

1 Reinforcement Learning Teachers of Test Time Scaling 22:37

1 [QA] Generalization or Hallucination? Understanding Out-of-Context Reasoning in Transformers 7:05

1 Generalization or Hallucination? Understanding Out-of-Context Reasoning in Transformers 18:28

1 [QA] Spurious Rewards: Rethinking Training Signals in RLVR 7:41

1 Spurious Rewards: Rethinking Training Signals in RLVR 30:13

1 [QA] Multiverse: Your Language Models Secretly Decide How to Parallelize and Merge Generation 8:08

1 Multiverse: Your Language Models Secretly Decide How to Parallelize and Merge Generation 24:30





1 [QA] Rewarding the Unlikely: Lifting GRPO Beyond Distribution Sharpening 7:45

1 Rewarding the Unlikely: Lifting GRPO Beyond Distribution Sharpening 16:56



1 [QA] Why Gradients Rapidly Increase Near the End of Training 7:00

1 Why Gradients Rapidly Increase Near the End of Training 11:24

1 [QA] GEM: Empowering LLM for both Embedding Generation and Language Understanding 7:41

1 GEM: Empowering LLM for both Embedding Generation and Language Understanding 20:38

1 [QA] HYPERSTEER: Activation Steering at Scale with Hypernetworks 7:49

1 HYPERSTEER: Activation Steering at Scale with Hypernetworks 9:15



1 [QA] Accelerating Diffusion LLMs via Adaptive Parallel Decoding 8:08

1 Accelerating Diffusion LLMs via Adaptive Parallel Decoding 21:09

1 [QA] Reflect, Retry, Reward: Self-Improving LLMs via Reinforcement Learning 7:34
Welcome to Player FM!
Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.