Every human being deserves to live their life to the fullest. Yet, every single day, drug addiction robs people and families of that opportunity. That problem is the driving force of this podcast. We provide a helping hand for people going through recovery; and a credible source for the truth about beating addiction. Every episode is a judgment-free space for healing, hope, and empowerment.
…
continue reading
AXRP (pronounced axe-urp) is the AI X-risk Research Podcast where I, Daniel Filan, have conversations with researchers about their papers. We discuss the paper, and hopefully get a sense of why it's been written and how it might reduce the risk of AI causing an existential catastrophe: that is, permanently and drastically curtailing humanity's future potential. You can visit the website and read transcripts at axrp.net.
…
continue reading
Your new favorite sports show. Tune in with Danache and special guests as she breaks down NFL hot topics, dissecting everything NFL on and off the field, game analysis, fantasy tips, game picks and more.
…
continue reading

1
40 - Jason Gross on Compact Proofs and Interpretability
2:36:05
2:36:05
Play later
Play later
Lists
Like
Liked
2:36:05How do we figure out whether interpretability is doing its job? One way is to see if it helps us prove things about models that we care about knowing. In this episode, I speak with Jason Gross about his agenda to benchmark interpretability in this way, and his exploration of the intersection of proofs and modern machine learning. Patreon: https://w…
…
continue reading

1
38.8 - David Duvenaud on Sabotage Evaluations and the Post-AGI Future
20:42
20:42
Play later
Play later
Lists
Like
Liked
20:42In this episode, I chat with David Duvenaud about two topics he's been thinking about: firstly, a paper he wrote about evaluating whether or not frontier models can sabotage human decision-making or monitoring of the same models; and secondly, the difficult situation humans find themselves in in a post-AGI future, even if AI is aligned with human i…
…
continue reading

1
38.7 - Anthony Aguirre on the Future of Life Institute
22:39
22:39
Play later
Play later
Lists
Like
Liked
22:39The Future of Life Institute is one of the oldest and most prominant organizations in the AI existential safety space, working on such topics as the AI pause open letter and how the EU AI Act can be improved. Metaculus is one of the premier forecasting sites on the internet. Behind both of them lie one man: Anthony Aguirre, who I talk with in this …
…
continue reading

1
38.6 - Joel Lehman on Positive Visions of AI
15:28
15:28
Play later
Play later
Lists
Like
Liked
15:28Typically this podcast talks about how to avert destruction from AI. But what would it take to ensure AI promotes human flourishing as well as it can? Is alignment to individuals enough, and if not, where do we go form here? In this episode, I talk with Joel Lehman about these questions. Patreon: https://www.patreon.com/axrpodcast Ko-fi: https://ko…
…
continue reading

1
38.5 - Adrià Garriga-Alonso on Detecting AI Scheming
27:41
27:41
Play later
Play later
Lists
Like
Liked
27:41Suppose we're worried about AIs engaging in long-term plans that they don't tell us about. If we were to peek inside their brains, what should we look for to check whether this was happening? In this episode Adrià Garriga-Alonso talks about his work trying to answer this question. Patreon: https://www.patreon.com/axrpodcast Ko-fi: https://ko-fi.com…
…
continue reading
AI researchers often complain about the poor coverage of their work in the news media. But why is this happening, and how can it be fixed? In this episode, I speak with Shakeel Hashim about the resource constraints facing AI journalism, the disconnect between journalists' and AI researchers' views on transformative AI, and efforts to improve the st…
…
continue reading
Lots of people in the AI safety space worry about models being able to make deliberate, multi-step plans. But can we already see this in existing neural nets? In this episode, I talk with Erik Jenner about his work looking at internal look-ahead within chess-playing neural networks. Patreon: https://www.patreon.com/axrpodcast Ko-fi: https://ko-fi.c…
…
continue reading

1
39 - Evan Hubinger on Model Organisms of Misalignment
1:45:47
1:45:47
Play later
Play later
Lists
Like
Liked
1:45:47The 'model organisms of misalignment' line of research creates AI models that exhibit various types of misalignment, and studies them to try to understand how the misalignment occurs and whether it can be somehow removed. In this episode, Evan Hubinger talks about two papers he's worked on at Anthropic under this agenda: "Sleeper Agents" and "Sycop…
…
continue reading

1
38.2 - Jesse Hoogland on Singular Learning Theory
18:18
18:18
Play later
Play later
Lists
Like
Liked
18:18You may have heard of singular learning theory, and its "local learning coefficient", or LLC - but have you heard of the refined LLC? In this episode, I chat with Jesse Hoogland about his work on SLT, and using the refined LLC to find a new circuit in language models. Patreon: https://www.patreon.com/axrpodcast Ko-fi: https://ko-fi.com/axrpodcast T…
…
continue reading
Road lines, street lights, and licence plates are examples of infrastructure used to ensure that roads operate smoothly. In this episode, Alan Chan talks about using similar interventions to help avoid bad outcomes from the deployment of AI agents. Patreon: https://www.patreon.com/axrpodcast Ko-fi: https://ko-fi.com/axrpodcast The transcript: https…
…
continue reading

1
38.0 - Zhijing Jin on LLMs, Causality, and Multi-Agent Systems
22:42
22:42
Play later
Play later
Lists
Like
Liked
22:42Do language models understand the causal structure of the world, or do they merely note correlations? And what happens when you build a big AI society out of them? In this brief episode, recorded at the Bay Area Alignment Workshop, I chat with Zhijing Jin about her research on these questions. Patreon: https://www.patreon.com/axrpodcast Ko-fi: http…
…
continue reading

1
37 - Jaime Sevilla on AI Forecasting
1:44:25
1:44:25
Play later
Play later
Lists
Like
Liked
1:44:25Epoch AI is the premier organization that tracks the trajectory of AI - how much compute is used, the role of algorithmic improvements, the growth in data used, and when the above trends might hit an end. In this episode, I speak with the director of Epoch AI, Jaime Sevilla, about how compute, data, and algorithmic improvements are impacting AI, an…
…
continue reading

1
36 - Adam Shai and Paul Riechers on Computational Mechanics
1:48:27
1:48:27
Play later
Play later
Lists
Like
Liked
1:48:27Sometimes, people talk about transformers as having "world models" as a result of being trained to predict text data on the internet. But what does this even mean? In this episode, I talk with Adam Shai and Paul Riechers about their work applying computational mechanics, a sub-field of physics studying how to predict random processes, to neural net…
…
continue reading
Patreon: https://www.patreon.com/axrpodcast MATS: https://www.matsprogram.org Note: I'm employed by MATS, but they're not paying me to make this video.
…
continue reading

1
35 - Peter Hase on LLM Beliefs and Easy-to-Hard Generalization
2:17:24
2:17:24
Play later
Play later
Lists
Like
Liked
2:17:24How do we figure out what large language models believe? In fact, do they even have beliefs? Do those beliefs have locations, and if so, can we edit those locations to change the beliefs? Also, how are we going to get AI to perform tasks so hard that we can't figure out if they succeeded at them? In this episode, I chat with Peter Hase about his re…
…
continue reading

1
34 - AI Evaluations with Beth Barnes
2:14:02
2:14:02
Play later
Play later
Lists
Like
Liked
2:14:02How can we figure out if AIs are capable enough to pose a threat to humans? When should we make a big effort to mitigate risks of catastrophic AI misbehaviour? In this episode, I chat with Beth Barnes, founder of and head of research at METR, about these questions and more. Patreon: patreon.com/axrpodcast Ko-fi: ko-fi.com/axrpodcast The transcript:…
…
continue reading

1
33 - RLHF Problems with Scott Emmons
1:41:24
1:41:24
Play later
Play later
Lists
Like
Liked
1:41:24Reinforcement Learning from Human Feedback, or RLHF, is one of the main ways that makers of large language models make them 'aligned'. But people have long noted that there are difficulties with this approach when the models are smarter than the humans providing feedback. In this episode, I talk with Scott Emmons about his work categorizing the pro…
…
continue reading

1
32 - Understanding Agency with Jan Kulveit
2:22:29
2:22:29
Play later
Play later
Lists
Like
Liked
2:22:29What's the difference between a large language model and the human brain? And what's wrong with our theories of agency? In this episode, I chat about these questions with Jan Kulveit, who leads the Alignment of Complex Systems research group. Patreon: patreon.com/axrpodcast Ko-fi: ko-fi.com/axrpodcast The transcript: axrp.net/episode/2024/05/30/epi…
…
continue reading

1
31 - Singular Learning Theory with Daniel Murfet
2:32:07
2:32:07
Play later
Play later
Lists
Like
Liked
2:32:07What's going on with deep learning? What sorts of models get learned, and what are the learning dynamics? Singular learning theory is a theory of Bayesian statistics broad enough in scope to encompass deep neural networks that may help answer these questions. In this episode, I speak with Daniel Murfet about this research program and what it tells …
…
continue reading

1
30 - AI Security with Jeffrey Ladish
2:15:44
2:15:44
Play later
Play later
Lists
Like
Liked
2:15:44Top labs use various forms of "safety training" on models before their release to make sure they don't do nasty stuff - but how robust is that? How can we ensure that the weights of powerful AIs don't get leaked or stolen? And what can AI even do these days? In this episode, I speak with Jeffrey Ladish about security and AI. Patreon: patreon.com/ax…
…
continue reading

1
29 - Science of Deep Learning with Vikrant Varma
2:13:46
2:13:46
Play later
Play later
Lists
Like
Liked
2:13:46In 2022, it was announced that a fairly simple method can be used to extract the true beliefs of a language model on any given topic, without having to actually understand the topic at hand. Earlier, in 2021, it was announced that neural networks sometimes 'grok': that is, when training them on certain tasks, they initially memorize their training …
…
continue reading

1
28 - Suing Labs for AI Risk with Gabriel Weil
1:57:30
1:57:30
Play later
Play later
Lists
Like
Liked
1:57:30How should the law govern AI? Those concerned about existential risks often push either for bans or for regulations meant to ensure that AI is developed safely - but another approach is possible. In this episode, Gabriel Weil talks about his proposal to modify tort law to enable people to sue AI companies for disasters that are "nearly catastrophic…
…
continue reading

1
27 - AI Control with Buck Shlegeris and Ryan Greenblatt
2:56:05
2:56:05
Play later
Play later
Lists
Like
Liked
2:56:05A lot of work to prevent AI existential risk takes the form of ensuring that AIs don't want to cause harm or take over the world---or in other words, ensuring that they're aligned. In this episode, I talk with Buck Shlegeris and Ryan Greenblatt about a different approach, called "AI control": ensuring that AI systems couldn't take over the world, e…
…
continue reading

1
26 - AI Governance with Elizabeth Seger
1:57:13
1:57:13
Play later
Play later
Lists
Like
Liked
1:57:13The events of this year have highlighted important questions about the governance of artificial intelligence. For instance, what does it mean to democratize AI? And how should we balance benefits and dangers of open-sourcing powerful AI systems such as large language models? In this episode, I speak with Elizabeth Seger about her research on these …
…
continue reading

1
25 - Cooperative AI with Caspar Oesterheld
3:02:09
3:02:09
Play later
Play later
Lists
Like
Liked
3:02:09Imagine a world where there are many powerful AI systems, working at cross purposes. You could suppose that different governments use AIs to manage their militaries, or simply that many powerful AIs have their own wills. At any rate, it seems valuable for them to be able to cooperatively work together and minimize pointless conflict. How do we ensu…
…
continue reading
Recently, OpenAI made a splash by announcing a new "Superalignment" team. Lead by Jan Leike and Ilya Sutskever, the team would consist of top researchers, attempting to solve alignment for superintelligent AIs in four years by figuring out how to build a trustworthy human-level AI alignment researcher, and then using it to solve the rest of the pro…
…
continue reading

1
23 - Mechanistic Anomaly Detection with Mark Xu
2:05:52
2:05:52
Play later
Play later
Lists
Like
Liked
2:05:52Is there some way we can detect bad behaviour in our AI system without having to know exactly what it looks like? In this episode, I speak with Mark Xu about mechanistic anomaly detection: a research direction based on the idea of detecting strange things happening in neural networks, in the hope that that will alert us of potential treacherous tur…
…
continue reading
Very brief survey: bit.ly/axrpsurvey2023 Store is closing in a week! Link: store.axrp.net/ Patreon: patreon.com/axrpodcast Ko-fi: ko-fi.com/axrpodcast
…
continue reading

1
22 - Shard Theory with Quintin Pope
3:28:21
3:28:21
Play later
Play later
Lists
Like
Liked
3:28:21What can we learn about advanced deep learning systems by understanding how humans learn and form values over their lifetimes? Will superhuman AI look like ruthless coherent utility optimization, or more like a mishmash of contextually activated desires? This episode's guest, Quintin Pope, has been thinking about these questions as a leading resear…
…
continue reading

1
21 - Interpretability for Engineers with Stephen Casper
1:56:02
1:56:02
Play later
Play later
Lists
Like
Liked
1:56:02Lots of people in the field of machine learning study 'interpretability', developing tools that they say give us useful information about neural networks. But how do we know if meaningful progress is actually being made? What should we want out of these tools? In this episode, I speak to Stephen Casper about these questions, as well as about a benc…
…
continue reading

1
20 - 'Reform' AI Alignment with Scott Aaronson
2:27:35
2:27:35
Play later
Play later
Lists
Like
Liked
2:27:35How should we scientifically think about the impact of AI on human civilization, and whether or not it will doom us all? In this episode, I speak with Scott Aaronson about his views on how to make progress in AI alignment, as well as his work on watermarking the output of language models, and how he moved from a background in quantum complexity the…
…
continue reading

1
Lions, the other teams, and Bears, Oh My!
42:10
42:10
Play later
Play later
Lists
Like
Liked
42:10Danache and Jerry rank their favorite off season moves so far. The Lamar Jackson Saga continues, AFC North needs and all the latest around the league.By Danache & Jeremy
…
continue reading
URP 2023 03 22By Danache & Jeremy
…
continue reading
Are you ready for some football? With the Jets on the market for a quarterback, it looks like Lamar Jackson's ball game! Dan Snyder, the Colts, the Titans, the Texans, the Bucks, and even the Patriots all need a quarterback. Who will be the one to take the lead? Tune in to find out!By Danache & Jeremy
…
continue reading
03-15-23By Danache & Jeremy
…
continue reading

1
MONEY, MONEY, MONEY w. Special Guest Lawrence "Bam Bam"
1:04:06
1:04:06
Play later
Play later
Lists
Like
Liked
1:04:06Tune is as the team discuss NFL combine, franchise tags, Derek Carr, Lamar Jackson, Aaron Rodgers, and break down what each NFL team needs, starting with the AFC East.By Danache & Jeremy
…
continue reading
This week Danache and Jerry talk about the lack of black coaches in the NFL, the NFL's unwillingness to hire Eric Bieniemy as a head coach after years of showing he's a capable OC, the Giants look like their heading in a mess of their own. Daniel Jones? Saquon? or both? and more..tune in.By Danache & Jeremy
…
continue reading
Danache and Jerry dive into Super Bowl LVII, the history and importance of the black quarterback, what a win or loss could mean for either Patrick Mahomes and Jalen Hurts. Are these two on the path to becoming one of the greatest to ever play the game? Andy Reid's legacy as a coach. Updates on some coaching moves around the league, LamarJackson and…
…
continue reading
Store: https://store.axrp.net/ Patreon: https://www.patreon.com/axrpodcast Ko-fi: https://ko-fi.com/axrpodcast Video: https://www.youtube.com/watch?v=kmPFjpEibu0
…
continue reading

1
19 - Mechanistic Interpretability with Neel Nanda
3:52:47
3:52:47
Play later
Play later
Lists
Like
Liked
3:52:47How good are we at understanding the internal computation of advanced machine learning models, and do we have a hope at getting better? In this episode, Neel Nanda talks about the sub-field of mechanistic interpretability research, as well as papers he's contributed to that explore the basics of transformer circuits, induction heads, and grokking. …
…
continue reading
Danache is joined by the Jospeh brothers and they break down what happened in the division championship games, the 49ers and their QB nightmares, coaching decisions that just make sense, Super Bowl preview and much more. And of course a heartfelt farewell to BradyBy Danache & Jeremy
…
continue reading
Danache and Jerry are joined by What Just Happened Sports founder Melissa Anthony. Tune in as they discuss, her journey, the importance of equipping and empowering women with sports knowledge, preview of this weekends conference championship games, and answering the question of what’s wrong with the Buffalo Bills?…
…
continue reading
Danache and Jerry talk super wild card weekend, Aaron Rodgers, does he stay, does he go? Lovie Smith and those good ol' Texans. Houston has a problem and it wasn't Lovie Smith. Game picks and more.By Danache & Jeremy
…
continue reading
Danache is back with special guests as the team discusses Buffalo Bills Safety Damar Hamlin, his road to recovery, the NFL community's next steps and more.By Danache & Jeremy
…
continue reading
Only thing better than 2 is 3. Jon, Jerry, & Ed talk Deion Sanders and the weight of the move. Are the Cowboys an OBJ away from best team in the league? The Bucs are all but guaranteed to make the playoffs but what's the point? The Bengals could be the most slept on team in the league and the 49ers might just be alright with Mr Irrelevant…
…
continue reading
The clock is ticking in the NFL & strength of schedule MATTERS. Jonathan & Jerry talk about the great differentiator between contenders and pretenders. You're QB can be YOUR guy, but not THE guy. The decisions that just have to be made, and welcome the fresh faces of the Coach of the Year race.By Danache & Jeremy
…
continue reading
Thanksgiving is here and we preview the best Thanksgiving line up we may have possibly seen in years. Zach Wilson benched..what are the JETS going to do? We might have an idea. Tune in.By Danache & Jeremy
…
continue reading
The NFL season is starting to turn up and the next few weeks can change everything. The AFC and NFC East have turned to probably the best divisons in football. Kirk Cousins isn't looking like a joke anymore (well 1 pm kirk that is). Tune is as we cover whats been going on around the league.By Danache & Jeremy
…
continue reading
Jon & Ed break down why Jeff Saturday makes sense and why it absolutely doesn't. Why the AFC East is the best division in the league. The biggest disappointments of the league and predictions for Week 10!By Danache & Jeremy
…
continue reading
The Jets & Giants are back. Did they get to t he party. a little too early or are they right on time and here to stay?By Danache & Jeremy
…
continue reading