<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet href="/stylesheet.xsl" type="text/xsl"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:podcast="https://podcastindex.org/namespace/1.0">
  <channel>
    <atom:link rel="self" type="application/rss+xml" href="https://feeds.transistor.fm/intoaisafety" title="MP3 Audio"/>
    <atom:link rel="hub" href="https://pubsubhubbub.appspot.com/"/>
    <podcast:podping usesPodping="true"/>
    <title>Into AI Safety</title>
    <generator>Transistor (https://transistor.fm)</generator>
    <itunes:new-feed-url>https://feeds.transistor.fm/intoaisafety</itunes:new-feed-url>
    <description>The Into AI Safety podcast aims to make it easier for everyone, regardless of background, to get meaningfully involved with the conversations surrounding the rules and regulations which should govern the research, development, deployment, and use of the technologies encompassed by the term "artificial intelligence" or "AI"

For better formatted show notes, additional resources, and more, go to https://kairos.fm/intoaisafety/</description>
    <copyright>© Kairos.fm</copyright>
    <podcast:guid>afd4d964-588e-5677-aac8-ef1ce12e6510</podcast:guid>
    <podcast:locked>yes</podcast:locked>
    <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
    <language>en-us</language>
    <pubDate>Wed, 15 Apr 2026 04:35:07 -0600</pubDate>
    <lastBuildDate>Wed, 15 Apr 2026 04:39:43 -0600</lastBuildDate>
    <link>https://kairos.fm/intoaisafety/</link>
    <image>
      <url>https://img.transistorcdn.com/t8Eetvpf3AgRbLLIjMLfZlGck8gh7daH1uKZWBFmlHI/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8yOTE5/MTU3M2U5NDAzYjJk/OTcwZTc5NzNkYzE3/ZWFjNC5qcGVn.jpg</url>
      <title>Into AI Safety</title>
      <link>https://kairos.fm/intoaisafety/</link>
    </image>
    <itunes:category text="Technology"/>
    <itunes:category text="Science">
      <itunes:category text="Mathematics"/>
    </itunes:category>
    <itunes:type>episodic</itunes:type>
    <itunes:author>Jacob Haimes</itunes:author>
    <itunes:image href="https://img.transistorcdn.com/t8Eetvpf3AgRbLLIjMLfZlGck8gh7daH1uKZWBFmlHI/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8yOTE5/MTU3M2U5NDAzYjJk/OTcwZTc5NzNkYzE3/ZWFjNC5qcGVn.jpg"/>
    <itunes:summary>The Into AI Safety podcast aims to make it easier for everyone, regardless of background, to get meaningfully involved with the conversations surrounding the rules and regulations which should govern the research, development, deployment, and use of the technologies encompassed by the term "artificial intelligence" or "AI"

For better formatted show notes, additional resources, and more, go to https://kairos.fm/intoaisafety/</itunes:summary>
    <itunes:subtitle>The Into AI Safety podcast aims to make it easier for everyone, regardless of background, to get meaningfully involved with the conversations surrounding the rules and regulations which should govern the research, development, deployment, and use of the technologies encompassed by the term "artificial intelligence" or "AI"

For better formatted show notes, additional resources, and more, go to https://kairos.fm/intoaisafety/.</itunes:subtitle>
    <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
    <itunes:owner>
      <itunes:name>Jacob Haimes</itunes:name>
      <itunes:email>listen@kairos.fm</itunes:email>
    </itunes:owner>
    <itunes:complete>No</itunes:complete>
    <itunes:explicit>No</itunes:explicit>
    <item>
      <title>Drawing Red Lines w/ Su Cizem</title>
      <itunes:episode>29</itunes:episode>
      <podcast:episode>29</podcast:episode>
      <itunes:title>Drawing Red Lines w/ Su Cizem</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">9be72f75-c464-4f82-9d67-11fd2f08dd45</guid>
      <link>https://kairos.fm/intoaisafety/e029</link>
      <description>
        <![CDATA[<p>Technology has been moving faster than policy for some time now, and the advent of AI isn't changing that, so what can we do to maintain safety despite uncertainty? Su Cizem has spent the last few years trying to answer that question. As an analyst at the Future Society, she works on global AI governance, specifically on building international consensus around AI red lines: the thresholds we collectively agree must never be crossed. In this conversation, Su walks through her path from philosophy to policy, the evolution of the global AI safety summit series, why voluntary commitments from AI labs aren't enough, and what it would actually take to make international cooperation on AI safety real.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Introduction</li>
<li>(03:23) - From Philosophy to Policy</li>
<li>(22:25) - What AI Governance Actually Means</li>
<li>(26:49) - The Summit Series</li>
<li>(43:01) - Drawing The Red Lines</li>
<li>(01:10:51) - Can These Companies Govern Themselves?</li>
<li>(01:24:01) - Breaking Into The Field</li>
<li>(01:27:51) - Closing Thoughts &amp; Outro</li>
</ul><br><strong>Critical Links<br></strong><em>Below are the most important links for this episode. For more, visit the episode </em><a href="https://kairos.fm/intoaisafety/e029"><em>page</em></a><em> on Kairos.fm.</em><ul><li>Su's <a href="https://www.linkedin.com/in/su-zeynep-cizem/">LinkedIn</a></li><li>Global <a href="https://red-lines.ai/">Call</a> for AI Red Lines</li><li>The Futures Society <a href="https://thefuturesociety.org/2025-athens-roundtable-report">report</a> - “Facing the Stakes of AI Together”: 2025 Athens Roundtable Report</li><li>Politico <a href="https://www.politico.eu/article/how-the-global-effort-to-keep-ai-safe-went-off-the-rails/">article</a> - How the global effort to keep AI safe went off the rails</li><li>TechPolicy.Press <a href="https://www.techpolicy.press/a-timeline-of-the-anthropic-pentagon-dispute/">article</a> - A Timeline of the Anthropic-Pentagon Dispute</li><li>The Guardian <a href="https://www.theguardian.com/news/2026/mar/26/ai-got-the-blame-for-the-iran-school-bombing-the-truth-is-far-more-worrying">article</a> - AI got the blame for the Iran school bombing. The truth is far more worrying</li><li>Google and OpenAI Employee <a href="https://notdivided.org/">open letter</a> - We Will Not Be Divided</li><li>The Register <a href="https://www.theregister.com/2026/03/06/openai_dod_deal/">article</a> - Altman said no to military AI abuses – then signed Pentagon deal anyway</li><li>SaferAI <a href="https://arxiv.org/html/2512.01166v3">report</a> - Evaluating AI Providers’ Frontier AI Safety Frameworks</li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Technology has been moving faster than policy for some time now, and the advent of AI isn't changing that, so what can we do to maintain safety despite uncertainty? Su Cizem has spent the last few years trying to answer that question. As an analyst at the Future Society, she works on global AI governance, specifically on building international consensus around AI red lines: the thresholds we collectively agree must never be crossed. In this conversation, Su walks through her path from philosophy to policy, the evolution of the global AI safety summit series, why voluntary commitments from AI labs aren't enough, and what it would actually take to make international cooperation on AI safety real.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Introduction</li>
<li>(03:23) - From Philosophy to Policy</li>
<li>(22:25) - What AI Governance Actually Means</li>
<li>(26:49) - The Summit Series</li>
<li>(43:01) - Drawing The Red Lines</li>
<li>(01:10:51) - Can These Companies Govern Themselves?</li>
<li>(01:24:01) - Breaking Into The Field</li>
<li>(01:27:51) - Closing Thoughts &amp; Outro</li>
</ul><br><strong>Critical Links<br></strong><em>Below are the most important links for this episode. For more, visit the episode </em><a href="https://kairos.fm/intoaisafety/e029"><em>page</em></a><em> on Kairos.fm.</em><ul><li>Su's <a href="https://www.linkedin.com/in/su-zeynep-cizem/">LinkedIn</a></li><li>Global <a href="https://red-lines.ai/">Call</a> for AI Red Lines</li><li>The Futures Society <a href="https://thefuturesociety.org/2025-athens-roundtable-report">report</a> - “Facing the Stakes of AI Together”: 2025 Athens Roundtable Report</li><li>Politico <a href="https://www.politico.eu/article/how-the-global-effort-to-keep-ai-safe-went-off-the-rails/">article</a> - How the global effort to keep AI safe went off the rails</li><li>TechPolicy.Press <a href="https://www.techpolicy.press/a-timeline-of-the-anthropic-pentagon-dispute/">article</a> - A Timeline of the Anthropic-Pentagon Dispute</li><li>The Guardian <a href="https://www.theguardian.com/news/2026/mar/26/ai-got-the-blame-for-the-iran-school-bombing-the-truth-is-far-more-worrying">article</a> - AI got the blame for the Iran school bombing. The truth is far more worrying</li><li>Google and OpenAI Employee <a href="https://notdivided.org/">open letter</a> - We Will Not Be Divided</li><li>The Register <a href="https://www.theregister.com/2026/03/06/openai_dod_deal/">article</a> - Altman said no to military AI abuses – then signed Pentagon deal anyway</li><li>SaferAI <a href="https://arxiv.org/html/2512.01166v3">report</a> - Evaluating AI Providers’ Frontier AI Safety Frameworks</li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 06 Apr 2026 10:30:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/1690daad/1ae6fb89.mp3" length="88141226" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/QGo42Vg8npuvDsxkhrlV9u1bglxIHB99v0vSvzXsrKg/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xODY1/MDBhNTFhMGFiZjJi/OGFjNjQ0M2Y3NzM1/MTE5Ni5wbmc.jpg"/>
      <itunes:duration>5507</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Technology has been moving faster than policy for some time now, and the advent of AI isn't changing that, so what can we do to maintain safety despite uncertainty? Su Cizem has spent the last few years trying to answer that question. As an analyst at the Future Society, she works on global AI governance, specifically on building international consensus around AI red lines: the thresholds we collectively agree must never be crossed. In this conversation, Su walks through her path from philosophy to policy, the evolution of the global AI safety summit series, why voluntary commitments from AI labs aren't enough, and what it would actually take to make international cooperation on AI safety real.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Introduction</li>
<li>(03:23) - From Philosophy to Policy</li>
<li>(22:25) - What AI Governance Actually Means</li>
<li>(26:49) - The Summit Series</li>
<li>(43:01) - Drawing The Red Lines</li>
<li>(01:10:51) - Can These Companies Govern Themselves?</li>
<li>(01:24:01) - Breaking Into The Field</li>
<li>(01:27:51) - Closing Thoughts &amp; Outro</li>
</ul><br><strong>Critical Links<br></strong><em>Below are the most important links for this episode. For more, visit the episode </em><a href="https://kairos.fm/intoaisafety/e029"><em>page</em></a><em> on Kairos.fm.</em><ul><li>Su's <a href="https://www.linkedin.com/in/su-zeynep-cizem/">LinkedIn</a></li><li>Global <a href="https://red-lines.ai/">Call</a> for AI Red Lines</li><li>The Futures Society <a href="https://thefuturesociety.org/2025-athens-roundtable-report">report</a> - “Facing the Stakes of AI Together”: 2025 Athens Roundtable Report</li><li>Politico <a href="https://www.politico.eu/article/how-the-global-effort-to-keep-ai-safe-went-off-the-rails/">article</a> - How the global effort to keep AI safe went off the rails</li><li>TechPolicy.Press <a href="https://www.techpolicy.press/a-timeline-of-the-anthropic-pentagon-dispute/">article</a> - A Timeline of the Anthropic-Pentagon Dispute</li><li>The Guardian <a href="https://www.theguardian.com/news/2026/mar/26/ai-got-the-blame-for-the-iran-school-bombing-the-truth-is-far-more-worrying">article</a> - AI got the blame for the Iran school bombing. The truth is far more worrying</li><li>Google and OpenAI Employee <a href="https://notdivided.org/">open letter</a> - We Will Not Be Divided</li><li>The Register <a href="https://www.theregister.com/2026/03/06/openai_dod_deal/">article</a> - Altman said no to military AI abuses – then signed Pentagon deal anyway</li><li>SaferAI <a href="https://arxiv.org/html/2512.01166v3">report</a> - Evaluating AI Providers’ Frontier AI Safety Frameworks</li></ul>]]>
      </itunes:summary>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Guest" href="https://intoaisafety.transistor.fm/people/su-cizem" img="https://img.transistorcdn.com/sGGxuorrC1zFkyFYCf0VaDyacgsOZ2qESzs_SGYrrLM/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS84MTAw/YzM1NmM3MjZiNGQ5/NjgyZTQ3Mjc2MDNh/MTcyMy5qcGVn.jpg">Su Cizem</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/1690daad/transcript.txt" type="text/plain"/>
      <podcast:chapters url="https://share.transistor.fm/s/1690daad/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Thinking Through "Digital Minds" w/ Jacy Reese-Anthis</title>
      <itunes:episode>28</itunes:episode>
      <podcast:episode>28</podcast:episode>
      <itunes:title>Thinking Through "Digital Minds" w/ Jacy Reese-Anthis</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">ed8de43f-63ce-452c-818b-e19a3c429779</guid>
      <link>https://kairos.fm/intoaisafety/e028/</link>
      <description>
        <![CDATA[<p>Jacy Reese-Anthis, founder of Sentience Institute and researcher at Stanford, began his journey working for animal welfare, but is now finishing up his PhD with research in many different AI subfields at the intersection of neuroscience, philosophy, social science, and machine learning. While this may seem like an odd jump at first, Jacy shares how his work has all been centered around the idea of moral circle expansion. In this episode, we dig into what sentience actually means (or at least how we can begin to think about it), why anthropomorphization is more complicated than it sounds, and how language models may be able to be leveraged as an effective tool for social science research.</p><p><br></p><p>Jacy also shares his median AGI estimate somewhere in there, so stay tuned if you want to catch it.</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><br><strong>Chapters</strong><br></p><ul><li>(00:00) - Introduction</li>
<li>(05:41) - From Animal Welfare to Digital Minds</li>
<li>(09:00) - Founding Sentience Institute</li>
<li>(22:00) - Defining Sentience</li>
<li>(27:13) - The Anthropomorphization Problem</li>
<li>(47:51) - Why "Digital Minds" (Not "Artificial Intelligence")</li>
<li>(51:05) - LLMs as Social Science Tools</li>
<li>(01:07:03) - Jacy’s AGI Timeline &amp; The Singularity</li>
<li>(01:09:23) - Final Thoughts &amp; Outro</li>
</ul><br><strong>Critical Links<br></strong><em>Below are the most important links for this episode. For more, visit the episode </em><a href="https://kairos.fm/intoaisafety/e028"><em>page</em></a><em> on Kairos.fm.</em><ul><li>Jacy's <a href="https://jacyanthis.com/">website</a></li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Jacy_Reese_Anthis">article</a> - Jacy Reese Anthis</li><li>Sentience Institute <a href="https://www.sentienceinstitute.org/">website</a></li><li>CHI <a href="https://arxiv.org/abs/2510.15905">paper</a> - Digital Companionship: Overlapping Uses of AI Companions and AI Assistants</li><li>ICML <a href="https://arxiv.org/abs/2504.02234">paper</a> - LLM Social Simulations Are a Promising Research Method</li><li>ACL <a href="https://arxiv.org/abs/2406.03198">paper</a> - The Impossibility of Fair LLMs</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/ELIZA_effect">article</a> - ELIZA effect</li><li>The Atlantic <a href="https://www.theatlantic.com/ideas/archive/2022/06/google-lamda-chatbot-sentient-ai/661322/">article</a> - How a Google Employee Fell for the Eliza Effect</li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Jacy Reese-Anthis, founder of Sentience Institute and researcher at Stanford, began his journey working for animal welfare, but is now finishing up his PhD with research in many different AI subfields at the intersection of neuroscience, philosophy, social science, and machine learning. While this may seem like an odd jump at first, Jacy shares how his work has all been centered around the idea of moral circle expansion. In this episode, we dig into what sentience actually means (or at least how we can begin to think about it), why anthropomorphization is more complicated than it sounds, and how language models may be able to be leveraged as an effective tool for social science research.</p><p><br></p><p>Jacy also shares his median AGI estimate somewhere in there, so stay tuned if you want to catch it.</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><br><strong>Chapters</strong><br></p><ul><li>(00:00) - Introduction</li>
<li>(05:41) - From Animal Welfare to Digital Minds</li>
<li>(09:00) - Founding Sentience Institute</li>
<li>(22:00) - Defining Sentience</li>
<li>(27:13) - The Anthropomorphization Problem</li>
<li>(47:51) - Why "Digital Minds" (Not "Artificial Intelligence")</li>
<li>(51:05) - LLMs as Social Science Tools</li>
<li>(01:07:03) - Jacy’s AGI Timeline &amp; The Singularity</li>
<li>(01:09:23) - Final Thoughts &amp; Outro</li>
</ul><br><strong>Critical Links<br></strong><em>Below are the most important links for this episode. For more, visit the episode </em><a href="https://kairos.fm/intoaisafety/e028"><em>page</em></a><em> on Kairos.fm.</em><ul><li>Jacy's <a href="https://jacyanthis.com/">website</a></li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Jacy_Reese_Anthis">article</a> - Jacy Reese Anthis</li><li>Sentience Institute <a href="https://www.sentienceinstitute.org/">website</a></li><li>CHI <a href="https://arxiv.org/abs/2510.15905">paper</a> - Digital Companionship: Overlapping Uses of AI Companions and AI Assistants</li><li>ICML <a href="https://arxiv.org/abs/2504.02234">paper</a> - LLM Social Simulations Are a Promising Research Method</li><li>ACL <a href="https://arxiv.org/abs/2406.03198">paper</a> - The Impossibility of Fair LLMs</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/ELIZA_effect">article</a> - ELIZA effect</li><li>The Atlantic <a href="https://www.theatlantic.com/ideas/archive/2022/06/google-lamda-chatbot-sentient-ai/661322/">article</a> - How a Google Employee Fell for the Eliza Effect</li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 09 Mar 2026 21:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/9a54c377/f3c949c8.mp3" length="136852745" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/eKbUGrrHJls2o6e6aHqR6g8dVv4i_mkYkMpdYgZ3-Zc/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9kMDk4/ODFmYmY3ZTZkMjA4/YTNlYWYxOGNmODQ3/ODExMy5wbmc.jpg"/>
      <itunes:duration>4276</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Jacy Reese-Anthis, founder of Sentience Institute and researcher at Stanford, began his journey working for animal welfare, but is now finishing up his PhD with research in many different AI subfields at the intersection of neuroscience, philosophy, social science, and machine learning. While this may seem like an odd jump at first, Jacy shares how his work has all been centered around the idea of moral circle expansion. In this episode, we dig into what sentience actually means (or at least how we can begin to think about it), why anthropomorphization is more complicated than it sounds, and how language models may be able to be leveraged as an effective tool for social science research.</p><p><br></p><p>Jacy also shares his median AGI estimate somewhere in there, so stay tuned if you want to catch it.</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><br><strong>Chapters</strong><br></p><ul><li>(00:00) - Introduction</li>
<li>(05:41) - From Animal Welfare to Digital Minds</li>
<li>(09:00) - Founding Sentience Institute</li>
<li>(22:00) - Defining Sentience</li>
<li>(27:13) - The Anthropomorphization Problem</li>
<li>(47:51) - Why "Digital Minds" (Not "Artificial Intelligence")</li>
<li>(51:05) - LLMs as Social Science Tools</li>
<li>(01:07:03) - Jacy’s AGI Timeline &amp; The Singularity</li>
<li>(01:09:23) - Final Thoughts &amp; Outro</li>
</ul><br><strong>Critical Links<br></strong><em>Below are the most important links for this episode. For more, visit the episode </em><a href="https://kairos.fm/intoaisafety/e028"><em>page</em></a><em> on Kairos.fm.</em><ul><li>Jacy's <a href="https://jacyanthis.com/">website</a></li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Jacy_Reese_Anthis">article</a> - Jacy Reese Anthis</li><li>Sentience Institute <a href="https://www.sentienceinstitute.org/">website</a></li><li>CHI <a href="https://arxiv.org/abs/2510.15905">paper</a> - Digital Companionship: Overlapping Uses of AI Companions and AI Assistants</li><li>ICML <a href="https://arxiv.org/abs/2504.02234">paper</a> - LLM Social Simulations Are a Promising Research Method</li><li>ACL <a href="https://arxiv.org/abs/2406.03198">paper</a> - The Impossibility of Fair LLMs</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/ELIZA_effect">article</a> - ELIZA effect</li><li>The Atlantic <a href="https://www.theatlantic.com/ideas/archive/2022/06/google-lamda-chatbot-sentient-ai/661322/">article</a> - How a Google Employee Fell for the Eliza Effect</li></ul>]]>
      </itunes:summary>
      <itunes:keywords>technology, AI, artificial intelligence, sentience, artificial sentience, artificial minds, AI anthropomorphization, digital minds, AI safety, AI welfare, AI companions, companion apps, perceptions of AI, sentience institute</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Guest" href="https://jacyanthis.com/" img="https://img.transistorcdn.com/w1_cnoo1tp6r4-YWyPnu-POUoQ7-9IgpjVbBomJ8eI4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8wZGQ2/N2NhYmFkYWRjYjQ2/YjM1NDBhMTk2ZGIy/YmUzNS53ZWJw.jpg">Jacy Reese-Anthis</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/9a54c377/transcript.txt" type="text/plain"/>
      <podcast:chapters url="https://share.transistor.fm/s/9a54c377/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Scaling AI Safety Through Mentorship w/ Dr. Ryan Kidd</title>
      <itunes:episode>27</itunes:episode>
      <podcast:episode>27</podcast:episode>
      <itunes:title>Scaling AI Safety Through Mentorship w/ Dr. Ryan Kidd</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">8fb864c3-29d4-4c26-b672-b9eb4710165c</guid>
      <link>https://kairos.fm/intoaisafety/e027</link>
      <description>
        <![CDATA[<p>What does it actually take to build a successful AI safety organization? I'm joined by Dr. Ryan Kidd, who has co-led MATS from a small pilot program to one of the field's premier talent pipelines. In this episode, he reveals the low-hanging fruit in AI safety field-building that most people are missing: the amplifier archetype.</p><p>I pushed Ryan on some hard questions, from balancing funder priorities and research independence, to building a robust selection process for both mentors and participants. Whether you're considering a career pivot into AI safety or already working in the field, this conversation offers practical advice on how to actually make an impact.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - - Intro</li>
<li>(08:16) - - Building MATS Post-FTX &amp; Summer of Love</li>
<li>(13:09) - - Balancing Funder Priorities and Research Independence</li>
<li>(19:44) - - The MATS Selection Process</li>
<li>(33:15) - - Talent Archetypes in AI Safety</li>
<li>(50:22) - - Comparative Advantage and Career Capital in AI Safety</li>
<li>(01:04:35) - - Building the AI Safety Ecosystem</li>
<li>(01:15:28) - - What Makes a Great AI Safety Amplifier</li>
<li>(01:21:44) - - Lightning Round Questions</li>
<li>(01:30:30) - - Final Thoughts &amp; Outro</li>
</ul><br><strong>Links</strong><ul><li><a href="https://matsprogram.org/apply?utm_source=ias&amp;utm_medium=&amp;utm_campaign=s26">MATS</a></li></ul><p><strong>Ryan's Writing</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/QzQQvGJYDeaDE4Cfg/talent-needs-of-technical-ai-safety-teams">post</a> - Talent needs of technical AI safety teams</li><li>LessWrong <a href="https://www.lesswrong.com/posts/yw9B5jQazBKGLjize/ai-safety-undervalues-founders">post</a> - AI safety undervalues founders</li><li>LessWrong <a href="https://www.lesswrong.com/posts/tPjAgWpsQrveFECWP/ryan-kidd-s-shortform?commentId=6dwq7qkszz5dyZ6sj">comment</a> - Comment permalink with 2025 MATS program details</li><li>LessWrong <a href="https://www.lesswrong.com/posts/WGNYAdBrsNjujJaB8/talk-ai-safety-fieldbuilding-at-mats">post</a> - Talk: AI Safety Fieldbuilding at MATS</li><li>LessWrong <a href="https://www.lesswrong.com/posts/LvswJts75fnpdjAEj/mats-mentor-selection">post</a> - MATS Mentor Selection</li><li>LessWrong <a href="https://www.lesswrong.com/posts/Yjiw5rnu4mJsYN8Xc/why-i-funded-pibbss-1">post</a> - Why I funded PIBBSS</li><li>EA Forum <a href="https://forum.effectivealtruism.org/posts/sGwPgwvaL2FkBHsRh/how-mats-addresses-mass-movement-building-concerns">post</a> - How MATS addresses mass movement building concerns</li></ul><p><strong>FTX Funding of AI Safety</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/WGpFFJo2uFe5ssgEb/an-overview-of-the-ai-safety-funding-situation">blogpost</a> - An Overview of the AI Safety Funding Situation</li><li>Fortune <a href="https://fortune.com/2022/11/15/sam-bankman-fried-ftx-collapse-a-i-safety-research-effective-altruism-debacle/">article</a> - Why Sam Bankman-Fried’s FTX debacle is roiling A.I. research</li><li>NY Times <a href="https://www.nytimes.com/2022/12/01/technology/sam-bankman-fried-crypto-artificial-intelligence.html">article</a> - FTX probes $6.5M in payments to AI safety group amid clawback crusade</li><li>Cointelegraph <a href="https://cointelegraph.com/news/crypto-exchange-ftx-subpoena-center-ai-safety-group-bankruptcy-proceedings">article</a> - FTX probes $6.5M in payments to AI safety group amid clawback crusade</li><li>FTX Future Fund <a href="https://archive.is/JYJJP">article</a> - Future Fund June 2022 Update (archive)</li><li>Tracxn <a href="https://tracxn.com/d/companies/anthropic/__SzoxXDMin-NK5tKB7ks8yHr6S9Mz68pjVCzFEcGFZ08/funding-and-investors#funding-rounds">page</a> - Anthropic Funding and Investors</li></ul><p><strong>Training &amp; Support Programs</strong></p><ul><li><a href="https://catalyze-impact.org/">Catalyze Impact</a></li><li><a href="https://seldonlab.com/">Seldon Lab</a></li><li><a href="https://sparai.org/">SPAR</a></li><li><a href="https://bluedot.org/">BlueDot Impact</a></li><li><a href="https://www.ycombinator.com/">YCombinator</a></li><li><a href="https://www.pivotal-research.org/">Pivotal</a></li><li><a href="https://researchathena.org/">Athena</a></li><li><a href="https://www.constellation.org/programs/astra-fellowship">Astra Fellowship</a></li><li><a href="https://horizonpublicservice.org/programs/become-a-fellow/">Horizon Fellowship</a></li><li><a href="https://www.baseresearch.org/base-fellowship">BASE Fellowship</a></li><li><a href="https://www.lasrlabs.org/">LASR Labs</a></li><li><a href="https://www.joinef.com/">Entrepeneur First</a></li></ul><p><strong>Funding Organizations</strong></p><ul><li><a href="https://coefficientgiving.org/">Coefficient Giving</a> (previously Open Philanthropy)</li><li><a href="https://funds.effectivealtruism.org/funds/far-future">LTFF</a></li><li><a href="https://www.longview.org/">Longview Philanthropy</a></li><li><a href="https://www.renaissancephilanthropy.org/">Renaissance Philanthropy</a></li></ul><p><strong>Coworking Spaces</strong></p><ul><li><a href="https://www.safeai.org.uk/">LISA</a></li><li><a href="https://moxsf.com/">Mox</a></li><li><a href="https://lighthaven.space/">Lighthaven</a></li><li><a href="https://www.far.ai/programs/far-labs">FAR Labs</a></li><li><a href="https://www.constellation.org/">Constellation</a></li><li><a href="https://collider.nyc/">Collider</a></li><li><a href="https://emergingthreat.net/">NET Office</a></li><li><a href="https://www.baish.com.ar/en">BAISH</a></li></ul><p><strong>Research Organizations &amp; Startups</strong></p><ul><li><a href="https://atla-ai.com/">Atla AI</a></li><li><a href="https://www.apolloresearch.ai/">Apollo Research</a></li><li><a href="https://timaeus.co/">Timaeus</a></li><li><a href="https://www.rand.org/global-and-emerging-risks/centers/ai-security-and-technology.html">RAND CAST</a></li><li><a href="https://humancompatible.ai/">CHAI</a></li></ul><p><strong>Other Sources</strong></p><ul><li>AXRP <a href="https://axrp.net/">website</a> - The AI X-risk Research Podcast</li><li>LessWrong <a href="https://www.lesswrong.com/posts/xqkGmfikqapbJ2YMj/shard-theory-an-overview">blogpost</a> - Shard Theory: An Overview</li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>What does it actually take to build a successful AI safety organization? I'm joined by Dr. Ryan Kidd, who has co-led MATS from a small pilot program to one of the field's premier talent pipelines. In this episode, he reveals the low-hanging fruit in AI safety field-building that most people are missing: the amplifier archetype.</p><p>I pushed Ryan on some hard questions, from balancing funder priorities and research independence, to building a robust selection process for both mentors and participants. Whether you're considering a career pivot into AI safety or already working in the field, this conversation offers practical advice on how to actually make an impact.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - - Intro</li>
<li>(08:16) - - Building MATS Post-FTX &amp; Summer of Love</li>
<li>(13:09) - - Balancing Funder Priorities and Research Independence</li>
<li>(19:44) - - The MATS Selection Process</li>
<li>(33:15) - - Talent Archetypes in AI Safety</li>
<li>(50:22) - - Comparative Advantage and Career Capital in AI Safety</li>
<li>(01:04:35) - - Building the AI Safety Ecosystem</li>
<li>(01:15:28) - - What Makes a Great AI Safety Amplifier</li>
<li>(01:21:44) - - Lightning Round Questions</li>
<li>(01:30:30) - - Final Thoughts &amp; Outro</li>
</ul><br><strong>Links</strong><ul><li><a href="https://matsprogram.org/apply?utm_source=ias&amp;utm_medium=&amp;utm_campaign=s26">MATS</a></li></ul><p><strong>Ryan's Writing</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/QzQQvGJYDeaDE4Cfg/talent-needs-of-technical-ai-safety-teams">post</a> - Talent needs of technical AI safety teams</li><li>LessWrong <a href="https://www.lesswrong.com/posts/yw9B5jQazBKGLjize/ai-safety-undervalues-founders">post</a> - AI safety undervalues founders</li><li>LessWrong <a href="https://www.lesswrong.com/posts/tPjAgWpsQrveFECWP/ryan-kidd-s-shortform?commentId=6dwq7qkszz5dyZ6sj">comment</a> - Comment permalink with 2025 MATS program details</li><li>LessWrong <a href="https://www.lesswrong.com/posts/WGNYAdBrsNjujJaB8/talk-ai-safety-fieldbuilding-at-mats">post</a> - Talk: AI Safety Fieldbuilding at MATS</li><li>LessWrong <a href="https://www.lesswrong.com/posts/LvswJts75fnpdjAEj/mats-mentor-selection">post</a> - MATS Mentor Selection</li><li>LessWrong <a href="https://www.lesswrong.com/posts/Yjiw5rnu4mJsYN8Xc/why-i-funded-pibbss-1">post</a> - Why I funded PIBBSS</li><li>EA Forum <a href="https://forum.effectivealtruism.org/posts/sGwPgwvaL2FkBHsRh/how-mats-addresses-mass-movement-building-concerns">post</a> - How MATS addresses mass movement building concerns</li></ul><p><strong>FTX Funding of AI Safety</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/WGpFFJo2uFe5ssgEb/an-overview-of-the-ai-safety-funding-situation">blogpost</a> - An Overview of the AI Safety Funding Situation</li><li>Fortune <a href="https://fortune.com/2022/11/15/sam-bankman-fried-ftx-collapse-a-i-safety-research-effective-altruism-debacle/">article</a> - Why Sam Bankman-Fried’s FTX debacle is roiling A.I. research</li><li>NY Times <a href="https://www.nytimes.com/2022/12/01/technology/sam-bankman-fried-crypto-artificial-intelligence.html">article</a> - FTX probes $6.5M in payments to AI safety group amid clawback crusade</li><li>Cointelegraph <a href="https://cointelegraph.com/news/crypto-exchange-ftx-subpoena-center-ai-safety-group-bankruptcy-proceedings">article</a> - FTX probes $6.5M in payments to AI safety group amid clawback crusade</li><li>FTX Future Fund <a href="https://archive.is/JYJJP">article</a> - Future Fund June 2022 Update (archive)</li><li>Tracxn <a href="https://tracxn.com/d/companies/anthropic/__SzoxXDMin-NK5tKB7ks8yHr6S9Mz68pjVCzFEcGFZ08/funding-and-investors#funding-rounds">page</a> - Anthropic Funding and Investors</li></ul><p><strong>Training &amp; Support Programs</strong></p><ul><li><a href="https://catalyze-impact.org/">Catalyze Impact</a></li><li><a href="https://seldonlab.com/">Seldon Lab</a></li><li><a href="https://sparai.org/">SPAR</a></li><li><a href="https://bluedot.org/">BlueDot Impact</a></li><li><a href="https://www.ycombinator.com/">YCombinator</a></li><li><a href="https://www.pivotal-research.org/">Pivotal</a></li><li><a href="https://researchathena.org/">Athena</a></li><li><a href="https://www.constellation.org/programs/astra-fellowship">Astra Fellowship</a></li><li><a href="https://horizonpublicservice.org/programs/become-a-fellow/">Horizon Fellowship</a></li><li><a href="https://www.baseresearch.org/base-fellowship">BASE Fellowship</a></li><li><a href="https://www.lasrlabs.org/">LASR Labs</a></li><li><a href="https://www.joinef.com/">Entrepeneur First</a></li></ul><p><strong>Funding Organizations</strong></p><ul><li><a href="https://coefficientgiving.org/">Coefficient Giving</a> (previously Open Philanthropy)</li><li><a href="https://funds.effectivealtruism.org/funds/far-future">LTFF</a></li><li><a href="https://www.longview.org/">Longview Philanthropy</a></li><li><a href="https://www.renaissancephilanthropy.org/">Renaissance Philanthropy</a></li></ul><p><strong>Coworking Spaces</strong></p><ul><li><a href="https://www.safeai.org.uk/">LISA</a></li><li><a href="https://moxsf.com/">Mox</a></li><li><a href="https://lighthaven.space/">Lighthaven</a></li><li><a href="https://www.far.ai/programs/far-labs">FAR Labs</a></li><li><a href="https://www.constellation.org/">Constellation</a></li><li><a href="https://collider.nyc/">Collider</a></li><li><a href="https://emergingthreat.net/">NET Office</a></li><li><a href="https://www.baish.com.ar/en">BAISH</a></li></ul><p><strong>Research Organizations &amp; Startups</strong></p><ul><li><a href="https://atla-ai.com/">Atla AI</a></li><li><a href="https://www.apolloresearch.ai/">Apollo Research</a></li><li><a href="https://timaeus.co/">Timaeus</a></li><li><a href="https://www.rand.org/global-and-emerging-risks/centers/ai-security-and-technology.html">RAND CAST</a></li><li><a href="https://humancompatible.ai/">CHAI</a></li></ul><p><strong>Other Sources</strong></p><ul><li>AXRP <a href="https://axrp.net/">website</a> - The AI X-risk Research Podcast</li><li>LessWrong <a href="https://www.lesswrong.com/posts/xqkGmfikqapbJ2YMj/shard-theory-an-overview">blogpost</a> - Shard Theory: An Overview</li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 02 Feb 2026 14:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/4b4ebed7/a4f61c90.mp3" length="175902975" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/tUb6iTKMppaaMFvz2wHbST2XDS4PZVmhGerXFgTYndw/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS82ZmNk/NzdhOGU5OTYzNjFj/ODRlOTY1NWVkOTY3/M2U2MC5wbmc.jpg"/>
      <itunes:duration>5496</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>What does it actually take to build a successful AI safety organization? I'm joined by Dr. Ryan Kidd, who has co-led MATS from a small pilot program to one of the field's premier talent pipelines. In this episode, he reveals the low-hanging fruit in AI safety field-building that most people are missing: the amplifier archetype.</p><p>I pushed Ryan on some hard questions, from balancing funder priorities and research independence, to building a robust selection process for both mentors and participants. Whether you're considering a career pivot into AI safety or already working in the field, this conversation offers practical advice on how to actually make an impact.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - - Intro</li>
<li>(08:16) - - Building MATS Post-FTX &amp; Summer of Love</li>
<li>(13:09) - - Balancing Funder Priorities and Research Independence</li>
<li>(19:44) - - The MATS Selection Process</li>
<li>(33:15) - - Talent Archetypes in AI Safety</li>
<li>(50:22) - - Comparative Advantage and Career Capital in AI Safety</li>
<li>(01:04:35) - - Building the AI Safety Ecosystem</li>
<li>(01:15:28) - - What Makes a Great AI Safety Amplifier</li>
<li>(01:21:44) - - Lightning Round Questions</li>
<li>(01:30:30) - - Final Thoughts &amp; Outro</li>
</ul><br><strong>Links</strong><ul><li><a href="https://matsprogram.org/apply?utm_source=ias&amp;utm_medium=&amp;utm_campaign=s26">MATS</a></li></ul><p><strong>Ryan's Writing</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/QzQQvGJYDeaDE4Cfg/talent-needs-of-technical-ai-safety-teams">post</a> - Talent needs of technical AI safety teams</li><li>LessWrong <a href="https://www.lesswrong.com/posts/yw9B5jQazBKGLjize/ai-safety-undervalues-founders">post</a> - AI safety undervalues founders</li><li>LessWrong <a href="https://www.lesswrong.com/posts/tPjAgWpsQrveFECWP/ryan-kidd-s-shortform?commentId=6dwq7qkszz5dyZ6sj">comment</a> - Comment permalink with 2025 MATS program details</li><li>LessWrong <a href="https://www.lesswrong.com/posts/WGNYAdBrsNjujJaB8/talk-ai-safety-fieldbuilding-at-mats">post</a> - Talk: AI Safety Fieldbuilding at MATS</li><li>LessWrong <a href="https://www.lesswrong.com/posts/LvswJts75fnpdjAEj/mats-mentor-selection">post</a> - MATS Mentor Selection</li><li>LessWrong <a href="https://www.lesswrong.com/posts/Yjiw5rnu4mJsYN8Xc/why-i-funded-pibbss-1">post</a> - Why I funded PIBBSS</li><li>EA Forum <a href="https://forum.effectivealtruism.org/posts/sGwPgwvaL2FkBHsRh/how-mats-addresses-mass-movement-building-concerns">post</a> - How MATS addresses mass movement building concerns</li></ul><p><strong>FTX Funding of AI Safety</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/WGpFFJo2uFe5ssgEb/an-overview-of-the-ai-safety-funding-situation">blogpost</a> - An Overview of the AI Safety Funding Situation</li><li>Fortune <a href="https://fortune.com/2022/11/15/sam-bankman-fried-ftx-collapse-a-i-safety-research-effective-altruism-debacle/">article</a> - Why Sam Bankman-Fried’s FTX debacle is roiling A.I. research</li><li>NY Times <a href="https://www.nytimes.com/2022/12/01/technology/sam-bankman-fried-crypto-artificial-intelligence.html">article</a> - FTX probes $6.5M in payments to AI safety group amid clawback crusade</li><li>Cointelegraph <a href="https://cointelegraph.com/news/crypto-exchange-ftx-subpoena-center-ai-safety-group-bankruptcy-proceedings">article</a> - FTX probes $6.5M in payments to AI safety group amid clawback crusade</li><li>FTX Future Fund <a href="https://archive.is/JYJJP">article</a> - Future Fund June 2022 Update (archive)</li><li>Tracxn <a href="https://tracxn.com/d/companies/anthropic/__SzoxXDMin-NK5tKB7ks8yHr6S9Mz68pjVCzFEcGFZ08/funding-and-investors#funding-rounds">page</a> - Anthropic Funding and Investors</li></ul><p><strong>Training &amp; Support Programs</strong></p><ul><li><a href="https://catalyze-impact.org/">Catalyze Impact</a></li><li><a href="https://seldonlab.com/">Seldon Lab</a></li><li><a href="https://sparai.org/">SPAR</a></li><li><a href="https://bluedot.org/">BlueDot Impact</a></li><li><a href="https://www.ycombinator.com/">YCombinator</a></li><li><a href="https://www.pivotal-research.org/">Pivotal</a></li><li><a href="https://researchathena.org/">Athena</a></li><li><a href="https://www.constellation.org/programs/astra-fellowship">Astra Fellowship</a></li><li><a href="https://horizonpublicservice.org/programs/become-a-fellow/">Horizon Fellowship</a></li><li><a href="https://www.baseresearch.org/base-fellowship">BASE Fellowship</a></li><li><a href="https://www.lasrlabs.org/">LASR Labs</a></li><li><a href="https://www.joinef.com/">Entrepeneur First</a></li></ul><p><strong>Funding Organizations</strong></p><ul><li><a href="https://coefficientgiving.org/">Coefficient Giving</a> (previously Open Philanthropy)</li><li><a href="https://funds.effectivealtruism.org/funds/far-future">LTFF</a></li><li><a href="https://www.longview.org/">Longview Philanthropy</a></li><li><a href="https://www.renaissancephilanthropy.org/">Renaissance Philanthropy</a></li></ul><p><strong>Coworking Spaces</strong></p><ul><li><a href="https://www.safeai.org.uk/">LISA</a></li><li><a href="https://moxsf.com/">Mox</a></li><li><a href="https://lighthaven.space/">Lighthaven</a></li><li><a href="https://www.far.ai/programs/far-labs">FAR Labs</a></li><li><a href="https://www.constellation.org/">Constellation</a></li><li><a href="https://collider.nyc/">Collider</a></li><li><a href="https://emergingthreat.net/">NET Office</a></li><li><a href="https://www.baish.com.ar/en">BAISH</a></li></ul><p><strong>Research Organizations &amp; Startups</strong></p><ul><li><a href="https://atla-ai.com/">Atla AI</a></li><li><a href="https://www.apolloresearch.ai/">Apollo Research</a></li><li><a href="https://timaeus.co/">Timaeus</a></li><li><a href="https://www.rand.org/global-and-emerging-risks/centers/ai-security-and-technology.html">RAND CAST</a></li><li><a href="https://humancompatible.ai/">CHAI</a></li></ul><p><strong>Other Sources</strong></p><ul><li>AXRP <a href="https://axrp.net/">website</a> - The AI X-risk Research Podcast</li><li>LessWrong <a href="https://www.lesswrong.com/posts/xqkGmfikqapbJ2YMj/shard-theory-an-overview">blogpost</a> - Shard Theory: An Overview</li></ul>]]>
      </itunes:summary>
      <itunes:keywords>artificial intelligence, technology, AI, AI safety, AI safety field-building, field building, mentorship, MATS, machine learning alignment and theory scholars, Ryan Kidd, AI security, upskilling, AI upskilling, Fellowship programs, AI fellowships, Open Philanthropy, Coefficient Giving, AI risk, AI safety research, SERI-MATS</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Guest" href="https://www.lesswrong.com/users/ryankidd44" img="https://img.transistorcdn.com/BcExI3YMvmS6Jn2Ey9TqLF_HOJpJCqGwaXK3mVDyKiM/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9iNjcw/OGQyZDk4OWQ1NzVm/ZDljNmYyYzkyY2Jm/OTg5Ni5qcGVn.jpg">Dr. Ryan Kidd</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/4b4ebed7/transcript.txt" type="text/plain"/>
      <podcast:chapters url="https://share.transistor.fm/s/4b4ebed7/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Sobering Up on AI Progress w/ Dr. Sean McGregor</title>
      <itunes:episode>26</itunes:episode>
      <podcast:episode>26</podcast:episode>
      <itunes:title>Sobering Up on AI Progress w/ Dr. Sean McGregor</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">b4b58807-63b7-4986-879e-8571ede6dc17</guid>
      <link>https://kairos.fm/intoaisafety/e026/</link>
      <description>
        <![CDATA[<p>Sean McGregor and I discuss about why evaluating AI systems has become so difficult; we cover everything from the breakdown of benchmarking, how incentives shape safety work, and what approaches like BenchRisk (his recent paper at NeurIPS) and AI auditing aim to fix as systems move into the real world. We also talk about his history and journey in AI safety, including his PhD on ML for public policy, how he started the AI Incident Database, and what he's working on now: AVERI, a non-profit for frontier model auditing.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Intro</li>
<li>(02:36) - What's broken about benchmarking</li>
<li>(03:41) - Sean’s wild PhD</li>
<li>(14:28) - The phantom internship</li>
<li>(19:25) - Sean's journey</li>
<li>(22:25) - Market-vs-regulatory modes and AIID</li>
<li>(32:13) - Drunk on AI progress</li>
<li>(38:34) - BenchRisk</li>
<li>(43:20) - Moral hazards and Master Hand</li>
<li>(50:34) - Liability, Section 230, and open source</li>
<li>(59:20) - AVERI</li>
<li>(01:11:30) - Closing thoughts &amp; outro</li>
</ul><br><strong>Links</strong><ul><li>Sean McGregor's <a href="https://seanbmcgregor.com/">website</a></li><li>AVERI <a href="https://www.averi.org/">website</a></li></ul><p><strong>BenchRisk</strong></p><ul><li>BenchRisk <a href="https://benchrisk.ai/">website</a></li><li>NeurIPS <a href="https://arxiv.org/abs/2510.21460">paper</a> - Risk Management for Mitigating Benchmark Failure Modes: BenchRisk</li><li>NeurIPS <a href="https://arxiv.org/abs/2111.15366">paper</a> - AI and the Everything in the Whole Wide World Benchmark</li></ul><p><strong>AIID</strong></p><ul><li>AI Incident Database <a href="https://incidentdatabase.ai/">website</a></li><li>IAAI <a href="https://arxiv.org/abs/2011.08512">paper</a> - Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database</li><li><a href="https://arxiv.org/pdf/2409.16425">Preprint</a> - Lessons for Editors of AI Incidents from the AI Incident Database</li><li>AIAAIC <a href="https://www.aiaaic.org/home">website</a> (another incident tracker)</li></ul><p><strong>Hot AI Summer</strong></p><ul><li>CACM <a href="https://homes.cs.washington.edu/~pedrod/papers/cacm12.pdf">article</a> - A Few Useful Things to Know About Machine Learning</li><li>CACM <a href="https://cacm.acm.org/opinion/how-the-ai-boom-went-bust/">article</a> - How the AI Boom Went Bust</li><li>Undergraduate <a href="https://www.researchgate.net/publication/333039347_Analyzing_the_Prospect_of_an_Approaching_AI_Winter">Thesis</a> - Analyzing the Prospect of an Approaching AI Winter</li><li>Tech Genies <a href="https://techgenies.com/ai-history-the-first-summer-and-winter-of-ai/">article</a> - AI History: The First Summer and Winter of AI</li><li>CACM <a href="https://cacm.acm.org/opinion/there-was-no-first-ai-winter/">article</a> - There Was No ‘First AI Winter’</li></ul><p><strong>Measuring Generalization</strong></p><ul><li>Neural Computation <a href="https://www.researchgate.net/publication/2755783_The_Lack_of_A_Priori_Distinctions_Between_Learning_Algorithms/link/54242c890cf238c6ea6e973c/download">article</a> - The Lack of A Priori Distinctions Between Learning Algorithms</li><li>ICLR <a href="https://arxiv.org/abs/1611.03530">paper</a> - Understanding deep learning requires rethinking generalization</li><li>ICML <a href="https://arxiv.org/abs/2305.01034v2">paper</a> - Model-agnostic Measure of Generalization Difficulty</li><li>Radiology Artificial Intelligence <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC9885377/">article</a> - Generalizability of Machine Learning Models: Quantitative Evaluation of Three Methodological Pitfalls</li><li><a href="https://arxiv.org/abs/2410.01769v2">Preprint</a> - Quantifying Generalization Complexity for Large Language Models</li></ul><p><strong>Insurers Exclude AI</strong></p><ul><li>Financial Times <a href="https://www.ft.com/content/abfe9741-f438-4ed6-a673-075ec177dc62">article</a> - Insurers retreat from AI cover as risk of multibillion-dollar claims mount</li><li>Tom's Hardware <a href="https://www.tomshardware.com/tech-industry/artificial-intelligence/insurers-move-to-limit-ai-liability-as-multi-billion-dollar-risks-emerge">article</a> - Major insurers move to avoid liability for AI lawsuits as multi-billion dollar risks emerge — Recent public incidents have lead to costly repercussions</li><li>Insurance Newsnet <a href="https://insurancenewsnet.com/oarticle/insurers-scale-back-ai-coverage-amid-fears-of-billion-dollar-claims">article</a> - Insurers Scale Back AI Coverage Amid Fears of Billion-Dollar Claims</li><li>Insurance Business <a href="https://www.insurancebusinessmag.com/us/news/technology/insurances-gen-ai-reckoning-has-come-559985.aspx">article</a> - Insurance’s gen AI reckoning has come</li></ul><p><strong>Section 230</strong></p><ul><li>Section 230 <a href="https://www.congress.gov/crs-product/R46751">overview</a></li><li>Legal <a href="https://www.congress.gov/crs-product/LSB11097">sidebar</a> - Section 230 Immunity and Generative Artificial Intelligence</li><li>Bad Internet Bills <a href="https://www.badinternetbills.com/">website</a></li><li>TechDirt <a href="https://www.techdirt.com/2025/12/17/section-230-faces-repeal-support-the-coverage-thats-been-getting-it-right-all-along/">article</a> - Section 230 Faces Repeal. Support The Coverage That’s Been Getting It Right All Along.</li><li>Privacy Guides <a href="https://www.privacyguides.org/videos/2025/12/16/taylor-lorenz-on-kosa-the-screen-act-and-repealing-section-230/">video</a> - Dissecting Bad Internet Bills with Taylor Lorenz: KOSA, SCREEN Act, Section 230</li><li>Journal of Technology in Behavioral Health <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC7785056/">article</a> - Social Media and Mental Health: Benefits, Risks, and Opportunities for Research and Practice</li><li>Time <a href="https://time.com/7337866/big-tech-social-media-regulation/">article</a> - Lawmakers Unveil New Bills to Curb Big Tech’s Power and Profit</li><li>House Hearing <a href="https://www.techpolicy.press/transcript-house-hearing-on-legislative-solutions-to-protect-children-and-teens-online/">transcript</a> - Legislative Solutions to Protect Children and Teens Online</li></ul><p><strong>Relevant Kairos.fm Episodes</strong></p><ul><li>Into AI Safety <a href="https://kairos.fm/intoaisafety/e023/">episode</a> - Growing BlueDot's Impact w/ Li-Lian Ang</li><li>muckrAIkers <a href="https://kairos.fm/muckraikers/e010/">episode</a> - NeurIPS 2024 Wrapped 🌯</li></ul><p><strong>Other Links</strong></p><ul><li>Encyclopedia of Life <a href="https://eol.org">website</a></li><li>IBM Watson AI XPRIZE <a href="https://www.xprize.org/competitions/artificial-intelligence">website</a></li><li>ML Commons <a href="https://mlcommons.org/">website</a></li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Enron_scandal">article</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Sean McGregor and I discuss about why evaluating AI systems has become so difficult; we cover everything from the breakdown of benchmarking, how incentives shape safety work, and what approaches like BenchRisk (his recent paper at NeurIPS) and AI auditing aim to fix as systems move into the real world. We also talk about his history and journey in AI safety, including his PhD on ML for public policy, how he started the AI Incident Database, and what he's working on now: AVERI, a non-profit for frontier model auditing.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Intro</li>
<li>(02:36) - What's broken about benchmarking</li>
<li>(03:41) - Sean’s wild PhD</li>
<li>(14:28) - The phantom internship</li>
<li>(19:25) - Sean's journey</li>
<li>(22:25) - Market-vs-regulatory modes and AIID</li>
<li>(32:13) - Drunk on AI progress</li>
<li>(38:34) - BenchRisk</li>
<li>(43:20) - Moral hazards and Master Hand</li>
<li>(50:34) - Liability, Section 230, and open source</li>
<li>(59:20) - AVERI</li>
<li>(01:11:30) - Closing thoughts &amp; outro</li>
</ul><br><strong>Links</strong><ul><li>Sean McGregor's <a href="https://seanbmcgregor.com/">website</a></li><li>AVERI <a href="https://www.averi.org/">website</a></li></ul><p><strong>BenchRisk</strong></p><ul><li>BenchRisk <a href="https://benchrisk.ai/">website</a></li><li>NeurIPS <a href="https://arxiv.org/abs/2510.21460">paper</a> - Risk Management for Mitigating Benchmark Failure Modes: BenchRisk</li><li>NeurIPS <a href="https://arxiv.org/abs/2111.15366">paper</a> - AI and the Everything in the Whole Wide World Benchmark</li></ul><p><strong>AIID</strong></p><ul><li>AI Incident Database <a href="https://incidentdatabase.ai/">website</a></li><li>IAAI <a href="https://arxiv.org/abs/2011.08512">paper</a> - Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database</li><li><a href="https://arxiv.org/pdf/2409.16425">Preprint</a> - Lessons for Editors of AI Incidents from the AI Incident Database</li><li>AIAAIC <a href="https://www.aiaaic.org/home">website</a> (another incident tracker)</li></ul><p><strong>Hot AI Summer</strong></p><ul><li>CACM <a href="https://homes.cs.washington.edu/~pedrod/papers/cacm12.pdf">article</a> - A Few Useful Things to Know About Machine Learning</li><li>CACM <a href="https://cacm.acm.org/opinion/how-the-ai-boom-went-bust/">article</a> - How the AI Boom Went Bust</li><li>Undergraduate <a href="https://www.researchgate.net/publication/333039347_Analyzing_the_Prospect_of_an_Approaching_AI_Winter">Thesis</a> - Analyzing the Prospect of an Approaching AI Winter</li><li>Tech Genies <a href="https://techgenies.com/ai-history-the-first-summer-and-winter-of-ai/">article</a> - AI History: The First Summer and Winter of AI</li><li>CACM <a href="https://cacm.acm.org/opinion/there-was-no-first-ai-winter/">article</a> - There Was No ‘First AI Winter’</li></ul><p><strong>Measuring Generalization</strong></p><ul><li>Neural Computation <a href="https://www.researchgate.net/publication/2755783_The_Lack_of_A_Priori_Distinctions_Between_Learning_Algorithms/link/54242c890cf238c6ea6e973c/download">article</a> - The Lack of A Priori Distinctions Between Learning Algorithms</li><li>ICLR <a href="https://arxiv.org/abs/1611.03530">paper</a> - Understanding deep learning requires rethinking generalization</li><li>ICML <a href="https://arxiv.org/abs/2305.01034v2">paper</a> - Model-agnostic Measure of Generalization Difficulty</li><li>Radiology Artificial Intelligence <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC9885377/">article</a> - Generalizability of Machine Learning Models: Quantitative Evaluation of Three Methodological Pitfalls</li><li><a href="https://arxiv.org/abs/2410.01769v2">Preprint</a> - Quantifying Generalization Complexity for Large Language Models</li></ul><p><strong>Insurers Exclude AI</strong></p><ul><li>Financial Times <a href="https://www.ft.com/content/abfe9741-f438-4ed6-a673-075ec177dc62">article</a> - Insurers retreat from AI cover as risk of multibillion-dollar claims mount</li><li>Tom's Hardware <a href="https://www.tomshardware.com/tech-industry/artificial-intelligence/insurers-move-to-limit-ai-liability-as-multi-billion-dollar-risks-emerge">article</a> - Major insurers move to avoid liability for AI lawsuits as multi-billion dollar risks emerge — Recent public incidents have lead to costly repercussions</li><li>Insurance Newsnet <a href="https://insurancenewsnet.com/oarticle/insurers-scale-back-ai-coverage-amid-fears-of-billion-dollar-claims">article</a> - Insurers Scale Back AI Coverage Amid Fears of Billion-Dollar Claims</li><li>Insurance Business <a href="https://www.insurancebusinessmag.com/us/news/technology/insurances-gen-ai-reckoning-has-come-559985.aspx">article</a> - Insurance’s gen AI reckoning has come</li></ul><p><strong>Section 230</strong></p><ul><li>Section 230 <a href="https://www.congress.gov/crs-product/R46751">overview</a></li><li>Legal <a href="https://www.congress.gov/crs-product/LSB11097">sidebar</a> - Section 230 Immunity and Generative Artificial Intelligence</li><li>Bad Internet Bills <a href="https://www.badinternetbills.com/">website</a></li><li>TechDirt <a href="https://www.techdirt.com/2025/12/17/section-230-faces-repeal-support-the-coverage-thats-been-getting-it-right-all-along/">article</a> - Section 230 Faces Repeal. Support The Coverage That’s Been Getting It Right All Along.</li><li>Privacy Guides <a href="https://www.privacyguides.org/videos/2025/12/16/taylor-lorenz-on-kosa-the-screen-act-and-repealing-section-230/">video</a> - Dissecting Bad Internet Bills with Taylor Lorenz: KOSA, SCREEN Act, Section 230</li><li>Journal of Technology in Behavioral Health <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC7785056/">article</a> - Social Media and Mental Health: Benefits, Risks, and Opportunities for Research and Practice</li><li>Time <a href="https://time.com/7337866/big-tech-social-media-regulation/">article</a> - Lawmakers Unveil New Bills to Curb Big Tech’s Power and Profit</li><li>House Hearing <a href="https://www.techpolicy.press/transcript-house-hearing-on-legislative-solutions-to-protect-children-and-teens-online/">transcript</a> - Legislative Solutions to Protect Children and Teens Online</li></ul><p><strong>Relevant Kairos.fm Episodes</strong></p><ul><li>Into AI Safety <a href="https://kairos.fm/intoaisafety/e023/">episode</a> - Growing BlueDot's Impact w/ Li-Lian Ang</li><li>muckrAIkers <a href="https://kairos.fm/muckraikers/e010/">episode</a> - NeurIPS 2024 Wrapped 🌯</li></ul><p><strong>Other Links</strong></p><ul><li>Encyclopedia of Life <a href="https://eol.org">website</a></li><li>IBM Watson AI XPRIZE <a href="https://www.xprize.org/competitions/artificial-intelligence">website</a></li><li>ML Commons <a href="https://mlcommons.org/">website</a></li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Enron_scandal">article</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 29 Dec 2025 13:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/8b720c1c/bdc3932f.mp3" length="70776411" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/TVBou2MsLVvZytwzRNQusmFDGOJ65dpXJXiHOE4yEpo/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZjEz/ZjgyNGRmMDIxYjQx/OTlmZjMxOTlhZjRl/NmEzMy5wbmc.jpg"/>
      <itunes:duration>4421</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Sean McGregor and I discuss about why evaluating AI systems has become so difficult; we cover everything from the breakdown of benchmarking, how incentives shape safety work, and what approaches like BenchRisk (his recent paper at NeurIPS) and AI auditing aim to fix as systems move into the real world. We also talk about his history and journey in AI safety, including his PhD on ML for public policy, how he started the AI Incident Database, and what he's working on now: AVERI, a non-profit for frontier model auditing.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Intro</li>
<li>(02:36) - What's broken about benchmarking</li>
<li>(03:41) - Sean’s wild PhD</li>
<li>(14:28) - The phantom internship</li>
<li>(19:25) - Sean's journey</li>
<li>(22:25) - Market-vs-regulatory modes and AIID</li>
<li>(32:13) - Drunk on AI progress</li>
<li>(38:34) - BenchRisk</li>
<li>(43:20) - Moral hazards and Master Hand</li>
<li>(50:34) - Liability, Section 230, and open source</li>
<li>(59:20) - AVERI</li>
<li>(01:11:30) - Closing thoughts &amp; outro</li>
</ul><br><strong>Links</strong><ul><li>Sean McGregor's <a href="https://seanbmcgregor.com/">website</a></li><li>AVERI <a href="https://www.averi.org/">website</a></li></ul><p><strong>BenchRisk</strong></p><ul><li>BenchRisk <a href="https://benchrisk.ai/">website</a></li><li>NeurIPS <a href="https://arxiv.org/abs/2510.21460">paper</a> - Risk Management for Mitigating Benchmark Failure Modes: BenchRisk</li><li>NeurIPS <a href="https://arxiv.org/abs/2111.15366">paper</a> - AI and the Everything in the Whole Wide World Benchmark</li></ul><p><strong>AIID</strong></p><ul><li>AI Incident Database <a href="https://incidentdatabase.ai/">website</a></li><li>IAAI <a href="https://arxiv.org/abs/2011.08512">paper</a> - Preventing Repeated Real World AI Failures by Cataloging Incidents: The AI Incident Database</li><li><a href="https://arxiv.org/pdf/2409.16425">Preprint</a> - Lessons for Editors of AI Incidents from the AI Incident Database</li><li>AIAAIC <a href="https://www.aiaaic.org/home">website</a> (another incident tracker)</li></ul><p><strong>Hot AI Summer</strong></p><ul><li>CACM <a href="https://homes.cs.washington.edu/~pedrod/papers/cacm12.pdf">article</a> - A Few Useful Things to Know About Machine Learning</li><li>CACM <a href="https://cacm.acm.org/opinion/how-the-ai-boom-went-bust/">article</a> - How the AI Boom Went Bust</li><li>Undergraduate <a href="https://www.researchgate.net/publication/333039347_Analyzing_the_Prospect_of_an_Approaching_AI_Winter">Thesis</a> - Analyzing the Prospect of an Approaching AI Winter</li><li>Tech Genies <a href="https://techgenies.com/ai-history-the-first-summer-and-winter-of-ai/">article</a> - AI History: The First Summer and Winter of AI</li><li>CACM <a href="https://cacm.acm.org/opinion/there-was-no-first-ai-winter/">article</a> - There Was No ‘First AI Winter’</li></ul><p><strong>Measuring Generalization</strong></p><ul><li>Neural Computation <a href="https://www.researchgate.net/publication/2755783_The_Lack_of_A_Priori_Distinctions_Between_Learning_Algorithms/link/54242c890cf238c6ea6e973c/download">article</a> - The Lack of A Priori Distinctions Between Learning Algorithms</li><li>ICLR <a href="https://arxiv.org/abs/1611.03530">paper</a> - Understanding deep learning requires rethinking generalization</li><li>ICML <a href="https://arxiv.org/abs/2305.01034v2">paper</a> - Model-agnostic Measure of Generalization Difficulty</li><li>Radiology Artificial Intelligence <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC9885377/">article</a> - Generalizability of Machine Learning Models: Quantitative Evaluation of Three Methodological Pitfalls</li><li><a href="https://arxiv.org/abs/2410.01769v2">Preprint</a> - Quantifying Generalization Complexity for Large Language Models</li></ul><p><strong>Insurers Exclude AI</strong></p><ul><li>Financial Times <a href="https://www.ft.com/content/abfe9741-f438-4ed6-a673-075ec177dc62">article</a> - Insurers retreat from AI cover as risk of multibillion-dollar claims mount</li><li>Tom's Hardware <a href="https://www.tomshardware.com/tech-industry/artificial-intelligence/insurers-move-to-limit-ai-liability-as-multi-billion-dollar-risks-emerge">article</a> - Major insurers move to avoid liability for AI lawsuits as multi-billion dollar risks emerge — Recent public incidents have lead to costly repercussions</li><li>Insurance Newsnet <a href="https://insurancenewsnet.com/oarticle/insurers-scale-back-ai-coverage-amid-fears-of-billion-dollar-claims">article</a> - Insurers Scale Back AI Coverage Amid Fears of Billion-Dollar Claims</li><li>Insurance Business <a href="https://www.insurancebusinessmag.com/us/news/technology/insurances-gen-ai-reckoning-has-come-559985.aspx">article</a> - Insurance’s gen AI reckoning has come</li></ul><p><strong>Section 230</strong></p><ul><li>Section 230 <a href="https://www.congress.gov/crs-product/R46751">overview</a></li><li>Legal <a href="https://www.congress.gov/crs-product/LSB11097">sidebar</a> - Section 230 Immunity and Generative Artificial Intelligence</li><li>Bad Internet Bills <a href="https://www.badinternetbills.com/">website</a></li><li>TechDirt <a href="https://www.techdirt.com/2025/12/17/section-230-faces-repeal-support-the-coverage-thats-been-getting-it-right-all-along/">article</a> - Section 230 Faces Repeal. Support The Coverage That’s Been Getting It Right All Along.</li><li>Privacy Guides <a href="https://www.privacyguides.org/videos/2025/12/16/taylor-lorenz-on-kosa-the-screen-act-and-repealing-section-230/">video</a> - Dissecting Bad Internet Bills with Taylor Lorenz: KOSA, SCREEN Act, Section 230</li><li>Journal of Technology in Behavioral Health <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC7785056/">article</a> - Social Media and Mental Health: Benefits, Risks, and Opportunities for Research and Practice</li><li>Time <a href="https://time.com/7337866/big-tech-social-media-regulation/">article</a> - Lawmakers Unveil New Bills to Curb Big Tech’s Power and Profit</li><li>House Hearing <a href="https://www.techpolicy.press/transcript-house-hearing-on-legislative-solutions-to-protect-children-and-teens-online/">transcript</a> - Legislative Solutions to Protect Children and Teens Online</li></ul><p><strong>Relevant Kairos.fm Episodes</strong></p><ul><li>Into AI Safety <a href="https://kairos.fm/intoaisafety/e023/">episode</a> - Growing BlueDot's Impact w/ Li-Lian Ang</li><li>muckrAIkers <a href="https://kairos.fm/muckraikers/e010/">episode</a> - NeurIPS 2024 Wrapped 🌯</li></ul><p><strong>Other Links</strong></p><ul><li>Encyclopedia of Life <a href="https://eol.org">website</a></li><li>IBM Watson AI XPRIZE <a href="https://www.xprize.org/competitions/artificial-intelligence">website</a></li><li>ML Commons <a href="https://mlcommons.org/">website</a></li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Enron_scandal">article</a></li></ul>]]>
      </itunes:summary>
      <itunes:keywords>AI, AI safety, technology, artificial intelligence, AI skepticism, AI hype, Benchmarking, AI Business, Machine Learning Safety, Generalization, Section 230, AIID, Sean McGregor, AVERI, LLM auditing, auditing LLMs, frontier model audits, bad internet bills, AI summer, BenchRisk, </itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://seanbmcgregor.com" img="https://img.transistorcdn.com/Ga2NtIp7Uug61_HIAdp-ukdUcARd6fPlFGqUDEvYlog/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS85MjYx/MWJhOWM0ZWY5ZDRi/MjAwODZmYzJiOWNm/OWJiMi5qcGc.jpg">Dr. Sean McGregor</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/8b720c1c/transcript.txt" type="text/plain"/>
      <podcast:chapters url="https://share.transistor.fm/s/8b720c1c/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Against 'The Singularity' w/ Dr. David Thorstad</title>
      <itunes:episode>25</itunes:episode>
      <podcast:episode>25</podcast:episode>
      <itunes:title>Against 'The Singularity' w/ Dr. David Thorstad</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">f3d8a985-e290-45ae-bda1-b8de2d57cc85</guid>
      <link>https://kairos.fm/intoaisafety/e025/</link>
      <description>
        <![CDATA[<p>Philosopher Dr. David Thorstad tears into one of AI safety's most influential arguments: the singularity hypothesis. We discuss why the idea of recursive self-improvement leading to superintelligence doesn't hold up under scrutiny, how these arguments have redirected hundreds of millions in funding away from proven interventions, and why people keep backpedaling to weaker versions when challenged.</p><p>David walks through the actual structure of singularity arguments, explains why similar patterns show up in other longtermist claims, and makes the case for why we should focus on concrete problems happening right now like poverty, disease, the rise of authoritarianism instead of speculative far-future scenarios.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Intro</li>
<li>(02:13) - David's background</li>
<li>(08:00) - (Against) The Singularity Hypothesis</li>
<li>(29:46) - Beyond the The Singularity</li>
<li>(39:56) - What We Should Actually Be Worried About</li>
<li>(49:00) - Philanthropic Funding</li>
</ul><br><strong>Links</strong><ul><li>David's personal <a href="https://www.dthorstad.com/">website</a></li><li><a href="https://reflectivealtruism.com/">Reflective Altruism</a>, David's blog</li></ul><p><strong>The Singularity Hypothesis</strong></p><ul><li>David's Philosophical Studies <a href="https://link.springer.com/article/10.1007/s11098-024-02143-5">article</a> - Against the singularity hypothesis</li><li>Time "AI Dictionary" <a href="https://time.com/collections/the-ai-dictionary-from-allbusiness-com/7273993/definition-of-singularity/">page</a> - Singularity</li><li>EA Forum <a href="https://forum.effectivealtruism.org/posts/N8CF4Nb7JHxRM2jKB/summary-against-the-singularity-hypothesis">blogpost</a> - Summary: Against the singularity hypothesis</li><li>Journal of Conciousness Studies <a href="https://gwern.net/doc/existential-risk/2016-chalmers.pdf">article</a> - The Singularity: A Philisophical Analysis</li><li>Interim <a href="https://link.springer.com/chapter/10.1007/978-3-642-32560-1_15#citeas">Report</a> from the Panel Chairs: AAAI Presidential Panel on Long-Term AI Futures</li><li>Epoch AI <a href="https://epoch.ai/blog/do-the-returns-to-software-rnd-point-towards-a-singularity">blogpost</a> - Do the returns to software R&amp;D point towards a singularity?</li><li>Epoch AI <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4814445">report</a> - Estimating Idea Production: A Methodological Survey</li></ul><p><strong>Funding References</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/WGpFFJo2uFe5ssgEb/an-overview-of-the-ai-safety-funding-situation">blogpost</a> - An Overview of the AI Safety Funding Situation</li><li>AISafety.com funding <a href="https://www.aisafety.com/funding">page</a></li><li><a href="https://hai.stanford.edu/assets/files/hai_ai_index_report_2025.pdf#page=248">Report</a> - Stanford AI Index 2025, Chapter 4.3</li><li>Forbes <a href="https://www.forbes.com/sites/bethkindig/2024/11/14/ai-spending-to-exceed-a-quarter-trillion-next-year/">article</a> - AI Spending To Exceed A Quarter Trillion Next Year</li><li>AI Panic <a href="https://www.aipanic.news/p/the-ai-existential-risk-industrial">article</a> - The “AI Existential Risk” Industrial Complex</li><li>GiveWell <a href="https://www.givewell.org/how-much-does-it-cost-to-save-a-life">webpage</a> - How Much Does It Cost To Save a Life?</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Purchasing_power_parity">article</a> - Purchasing power parity</li></ul><p><strong>Pascal's Mugging and the St. Petersburg Paradox</strong></p><ul><li>Wikipedia <a href="https://en.wikipedia.org/wiki/St._Petersburg_paradox">article</a> - St. Petersburg Paradox</li><li>Conjecture Magazine <a href="https://medium.com/conjecture-magazine/pascals-mugging-and-the-poverty-of-the-expected-value-calculus-70b190d953cd">article</a> - Pascal’s Mugging and Bad Explanations</li><li>neurabites <a href="https://neurabites.com/ergodicity/">explainer</a> - Ergodicity: the Most Over-Looked Assumption</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Extraordinary_claims_require_extraordinary_evidence">article</a> - Extraordinary claims require extraordinary evidence</li></ul><p><strong>The Time of Perils</strong></p><ul><li>Global Priorities Institute <a href="https://philarchive.org/rec/THOERP">working paper</a> - Existential risk pessimism and the time of perils</li><li>Ethics <a href="https://www.journals.uchicago.edu/doi/abs/10.1086/731436?journalCode=et">article</a> - Mistakes in the Moral Mathematics of Existential Risk</li><li>Philosophy &amp; Public Affairs <a href="https://onlinelibrary.wiley.com/doi/10.1111/papa.12248">article</a> - High Risk, Low Reward: A Challenge to the Astronomical Value of Existential Risk Mitigation</li><li>Toby Ord <a href="https://theprecipice.com/">Book</a> - The Precipice</li><li>Rethink Priorities <a href="https://rethinkpriorities.org/research-area/charting-the-precipice-the-time-of-perils-and-prioritizing-x-risk/">blogpost</a> - Charting the precipice</li><li>AI Futures Project <a href="https://ai-2027.com/">blogpost</a> - AI 2027</li></ul><p><strong>Trump's Higher Education </strong><strong>Threat</strong><strong> Compact</strong></p><ul><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Compact_for_Academic_Excellence_in_Higher_Education">article</a> - Compact for Academic Excellence in Higher Education</li><li>Pen America <a href="https://pen.org/trumps-compact-for-higher-education-faq/">explainer</a> - What is Trump’s Compact for Higher Education? And More Frequently Asked Questions</li><li><a href="https://vanderbiltaaup.org/2025/10/06/statement-by-the-vanderbilt-aaup-executive-committee-on-the-compact-for-academic-excellence-in-higher-education/">Statement</a> by the Vanderbilt AAUP Executive Committee on the “Compact for Academic Excellence in Higher Education”</li><li>The Vanderbilt Hustler <a href="https://vanderbilthustler.com/2025/10/20/breaking-chancellor-daniel-diermeier-fails-to-reject-higher-education-compact-reaffirms-vanderbilts-values-and-openness-to-discussion/">article</a> - BREAKING: Chancellor Daniel Diermeier fails to reject higher education compact, reaffirms Vanderbilt’s values and openness to discussion</li><li>The Vanderbilt Hustler <a href="https://vanderbilthustler.com/2025/11/06/students-and-faculty-organize-rally-outside-kirkland-hall-against-trump-administrations-higher-education-compact/">article</a> - Students and faculty organize rally outside Kirkland Hall against Trump administration’s higher education compact</li><li>Free Speech Center <a href="https://firstamendment.mtsu.edu/article/compact-for-academic-excellence/">article</a> - Compact for Academic Excellence</li></ul><p><strong>More of David's Work</strong></p><ul><li>Global Priorities Institute <a href="https://www.globalprioritiesinstitute.org/wp-content/uploads/David-Thorstad-What-power-seeking-theorems-do-not-show.pdf">working paper</a> - What power-seeking theorems do not show</li><li><a href="https://global.oup.com/academic/product/essays-on-longtermism-9780192883858?cc=us&amp;lang=en&amp;">Book</a> - Essays on Longtermism</li></ul><p><strong>Vibe Shift</strong></p><ul><li>Blood in the Machine <a href="https://www.bloodinthemachine.com/p/gpt-5-is-a-joke-will-it-matter">article</a> - GPT-5 Is a Joke. Will It Matter?</li><li>Futurism <a href="https://futurism.com/gpt-5-underwhelming">article</a> - Evidence Grows That GPT-5 Is a Bit of a Dud</li><li>Gary Marcus <a href="https://garymarcus.substack.com/p/gpt-5-overdue-overhyped-and-underwhelming">substack</a> - GPT-5: Overdue, overhyped and underwhelming. And that’s not the worst of it.</li><li>Pew Research <a href="https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/">report</a> - How the U.S. Public and AI Experts View Artificial Intelligence</li><li>N...</li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Philosopher Dr. David Thorstad tears into one of AI safety's most influential arguments: the singularity hypothesis. We discuss why the idea of recursive self-improvement leading to superintelligence doesn't hold up under scrutiny, how these arguments have redirected hundreds of millions in funding away from proven interventions, and why people keep backpedaling to weaker versions when challenged.</p><p>David walks through the actual structure of singularity arguments, explains why similar patterns show up in other longtermist claims, and makes the case for why we should focus on concrete problems happening right now like poverty, disease, the rise of authoritarianism instead of speculative far-future scenarios.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Intro</li>
<li>(02:13) - David's background</li>
<li>(08:00) - (Against) The Singularity Hypothesis</li>
<li>(29:46) - Beyond the The Singularity</li>
<li>(39:56) - What We Should Actually Be Worried About</li>
<li>(49:00) - Philanthropic Funding</li>
</ul><br><strong>Links</strong><ul><li>David's personal <a href="https://www.dthorstad.com/">website</a></li><li><a href="https://reflectivealtruism.com/">Reflective Altruism</a>, David's blog</li></ul><p><strong>The Singularity Hypothesis</strong></p><ul><li>David's Philosophical Studies <a href="https://link.springer.com/article/10.1007/s11098-024-02143-5">article</a> - Against the singularity hypothesis</li><li>Time "AI Dictionary" <a href="https://time.com/collections/the-ai-dictionary-from-allbusiness-com/7273993/definition-of-singularity/">page</a> - Singularity</li><li>EA Forum <a href="https://forum.effectivealtruism.org/posts/N8CF4Nb7JHxRM2jKB/summary-against-the-singularity-hypothesis">blogpost</a> - Summary: Against the singularity hypothesis</li><li>Journal of Conciousness Studies <a href="https://gwern.net/doc/existential-risk/2016-chalmers.pdf">article</a> - The Singularity: A Philisophical Analysis</li><li>Interim <a href="https://link.springer.com/chapter/10.1007/978-3-642-32560-1_15#citeas">Report</a> from the Panel Chairs: AAAI Presidential Panel on Long-Term AI Futures</li><li>Epoch AI <a href="https://epoch.ai/blog/do-the-returns-to-software-rnd-point-towards-a-singularity">blogpost</a> - Do the returns to software R&amp;D point towards a singularity?</li><li>Epoch AI <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4814445">report</a> - Estimating Idea Production: A Methodological Survey</li></ul><p><strong>Funding References</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/WGpFFJo2uFe5ssgEb/an-overview-of-the-ai-safety-funding-situation">blogpost</a> - An Overview of the AI Safety Funding Situation</li><li>AISafety.com funding <a href="https://www.aisafety.com/funding">page</a></li><li><a href="https://hai.stanford.edu/assets/files/hai_ai_index_report_2025.pdf#page=248">Report</a> - Stanford AI Index 2025, Chapter 4.3</li><li>Forbes <a href="https://www.forbes.com/sites/bethkindig/2024/11/14/ai-spending-to-exceed-a-quarter-trillion-next-year/">article</a> - AI Spending To Exceed A Quarter Trillion Next Year</li><li>AI Panic <a href="https://www.aipanic.news/p/the-ai-existential-risk-industrial">article</a> - The “AI Existential Risk” Industrial Complex</li><li>GiveWell <a href="https://www.givewell.org/how-much-does-it-cost-to-save-a-life">webpage</a> - How Much Does It Cost To Save a Life?</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Purchasing_power_parity">article</a> - Purchasing power parity</li></ul><p><strong>Pascal's Mugging and the St. Petersburg Paradox</strong></p><ul><li>Wikipedia <a href="https://en.wikipedia.org/wiki/St._Petersburg_paradox">article</a> - St. Petersburg Paradox</li><li>Conjecture Magazine <a href="https://medium.com/conjecture-magazine/pascals-mugging-and-the-poverty-of-the-expected-value-calculus-70b190d953cd">article</a> - Pascal’s Mugging and Bad Explanations</li><li>neurabites <a href="https://neurabites.com/ergodicity/">explainer</a> - Ergodicity: the Most Over-Looked Assumption</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Extraordinary_claims_require_extraordinary_evidence">article</a> - Extraordinary claims require extraordinary evidence</li></ul><p><strong>The Time of Perils</strong></p><ul><li>Global Priorities Institute <a href="https://philarchive.org/rec/THOERP">working paper</a> - Existential risk pessimism and the time of perils</li><li>Ethics <a href="https://www.journals.uchicago.edu/doi/abs/10.1086/731436?journalCode=et">article</a> - Mistakes in the Moral Mathematics of Existential Risk</li><li>Philosophy &amp; Public Affairs <a href="https://onlinelibrary.wiley.com/doi/10.1111/papa.12248">article</a> - High Risk, Low Reward: A Challenge to the Astronomical Value of Existential Risk Mitigation</li><li>Toby Ord <a href="https://theprecipice.com/">Book</a> - The Precipice</li><li>Rethink Priorities <a href="https://rethinkpriorities.org/research-area/charting-the-precipice-the-time-of-perils-and-prioritizing-x-risk/">blogpost</a> - Charting the precipice</li><li>AI Futures Project <a href="https://ai-2027.com/">blogpost</a> - AI 2027</li></ul><p><strong>Trump's Higher Education </strong><strong>Threat</strong><strong> Compact</strong></p><ul><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Compact_for_Academic_Excellence_in_Higher_Education">article</a> - Compact for Academic Excellence in Higher Education</li><li>Pen America <a href="https://pen.org/trumps-compact-for-higher-education-faq/">explainer</a> - What is Trump’s Compact for Higher Education? And More Frequently Asked Questions</li><li><a href="https://vanderbiltaaup.org/2025/10/06/statement-by-the-vanderbilt-aaup-executive-committee-on-the-compact-for-academic-excellence-in-higher-education/">Statement</a> by the Vanderbilt AAUP Executive Committee on the “Compact for Academic Excellence in Higher Education”</li><li>The Vanderbilt Hustler <a href="https://vanderbilthustler.com/2025/10/20/breaking-chancellor-daniel-diermeier-fails-to-reject-higher-education-compact-reaffirms-vanderbilts-values-and-openness-to-discussion/">article</a> - BREAKING: Chancellor Daniel Diermeier fails to reject higher education compact, reaffirms Vanderbilt’s values and openness to discussion</li><li>The Vanderbilt Hustler <a href="https://vanderbilthustler.com/2025/11/06/students-and-faculty-organize-rally-outside-kirkland-hall-against-trump-administrations-higher-education-compact/">article</a> - Students and faculty organize rally outside Kirkland Hall against Trump administration’s higher education compact</li><li>Free Speech Center <a href="https://firstamendment.mtsu.edu/article/compact-for-academic-excellence/">article</a> - Compact for Academic Excellence</li></ul><p><strong>More of David's Work</strong></p><ul><li>Global Priorities Institute <a href="https://www.globalprioritiesinstitute.org/wp-content/uploads/David-Thorstad-What-power-seeking-theorems-do-not-show.pdf">working paper</a> - What power-seeking theorems do not show</li><li><a href="https://global.oup.com/academic/product/essays-on-longtermism-9780192883858?cc=us&amp;lang=en&amp;">Book</a> - Essays on Longtermism</li></ul><p><strong>Vibe Shift</strong></p><ul><li>Blood in the Machine <a href="https://www.bloodinthemachine.com/p/gpt-5-is-a-joke-will-it-matter">article</a> - GPT-5 Is a Joke. Will It Matter?</li><li>Futurism <a href="https://futurism.com/gpt-5-underwhelming">article</a> - Evidence Grows That GPT-5 Is a Bit of a Dud</li><li>Gary Marcus <a href="https://garymarcus.substack.com/p/gpt-5-overdue-overhyped-and-underwhelming">substack</a> - GPT-5: Overdue, overhyped and underwhelming. And that’s not the worst of it.</li><li>Pew Research <a href="https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/">report</a> - How the U.S. Public and AI Experts View Artificial Intelligence</li><li>N...</li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 24 Nov 2025 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/5c0f0e45/6f561636.mp3" length="66446723" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/9vWfsSMBGpHMqH4I6btk4T_MaG82tD1EpibLA1uY3tk/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS82YTA4/ODdkYzJjN2EzZDE0/MTZjMmM1MzQ5N2Ni/MDA1Zi5wbmc.jpg"/>
      <itunes:duration>4150</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Philosopher Dr. David Thorstad tears into one of AI safety's most influential arguments: the singularity hypothesis. We discuss why the idea of recursive self-improvement leading to superintelligence doesn't hold up under scrutiny, how these arguments have redirected hundreds of millions in funding away from proven interventions, and why people keep backpedaling to weaker versions when challenged.</p><p>David walks through the actual structure of singularity arguments, explains why similar patterns show up in other longtermist claims, and makes the case for why we should focus on concrete problems happening right now like poverty, disease, the rise of authoritarianism instead of speculative far-future scenarios.</p><p><strong>Chapters</strong><br></p><ul><li>(00:00) - Intro</li>
<li>(02:13) - David's background</li>
<li>(08:00) - (Against) The Singularity Hypothesis</li>
<li>(29:46) - Beyond the The Singularity</li>
<li>(39:56) - What We Should Actually Be Worried About</li>
<li>(49:00) - Philanthropic Funding</li>
</ul><br><strong>Links</strong><ul><li>David's personal <a href="https://www.dthorstad.com/">website</a></li><li><a href="https://reflectivealtruism.com/">Reflective Altruism</a>, David's blog</li></ul><p><strong>The Singularity Hypothesis</strong></p><ul><li>David's Philosophical Studies <a href="https://link.springer.com/article/10.1007/s11098-024-02143-5">article</a> - Against the singularity hypothesis</li><li>Time "AI Dictionary" <a href="https://time.com/collections/the-ai-dictionary-from-allbusiness-com/7273993/definition-of-singularity/">page</a> - Singularity</li><li>EA Forum <a href="https://forum.effectivealtruism.org/posts/N8CF4Nb7JHxRM2jKB/summary-against-the-singularity-hypothesis">blogpost</a> - Summary: Against the singularity hypothesis</li><li>Journal of Conciousness Studies <a href="https://gwern.net/doc/existential-risk/2016-chalmers.pdf">article</a> - The Singularity: A Philisophical Analysis</li><li>Interim <a href="https://link.springer.com/chapter/10.1007/978-3-642-32560-1_15#citeas">Report</a> from the Panel Chairs: AAAI Presidential Panel on Long-Term AI Futures</li><li>Epoch AI <a href="https://epoch.ai/blog/do-the-returns-to-software-rnd-point-towards-a-singularity">blogpost</a> - Do the returns to software R&amp;D point towards a singularity?</li><li>Epoch AI <a href="https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4814445">report</a> - Estimating Idea Production: A Methodological Survey</li></ul><p><strong>Funding References</strong></p><ul><li>LessWrong <a href="https://www.lesswrong.com/posts/WGpFFJo2uFe5ssgEb/an-overview-of-the-ai-safety-funding-situation">blogpost</a> - An Overview of the AI Safety Funding Situation</li><li>AISafety.com funding <a href="https://www.aisafety.com/funding">page</a></li><li><a href="https://hai.stanford.edu/assets/files/hai_ai_index_report_2025.pdf#page=248">Report</a> - Stanford AI Index 2025, Chapter 4.3</li><li>Forbes <a href="https://www.forbes.com/sites/bethkindig/2024/11/14/ai-spending-to-exceed-a-quarter-trillion-next-year/">article</a> - AI Spending To Exceed A Quarter Trillion Next Year</li><li>AI Panic <a href="https://www.aipanic.news/p/the-ai-existential-risk-industrial">article</a> - The “AI Existential Risk” Industrial Complex</li><li>GiveWell <a href="https://www.givewell.org/how-much-does-it-cost-to-save-a-life">webpage</a> - How Much Does It Cost To Save a Life?</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Purchasing_power_parity">article</a> - Purchasing power parity</li></ul><p><strong>Pascal's Mugging and the St. Petersburg Paradox</strong></p><ul><li>Wikipedia <a href="https://en.wikipedia.org/wiki/St._Petersburg_paradox">article</a> - St. Petersburg Paradox</li><li>Conjecture Magazine <a href="https://medium.com/conjecture-magazine/pascals-mugging-and-the-poverty-of-the-expected-value-calculus-70b190d953cd">article</a> - Pascal’s Mugging and Bad Explanations</li><li>neurabites <a href="https://neurabites.com/ergodicity/">explainer</a> - Ergodicity: the Most Over-Looked Assumption</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Extraordinary_claims_require_extraordinary_evidence">article</a> - Extraordinary claims require extraordinary evidence</li></ul><p><strong>The Time of Perils</strong></p><ul><li>Global Priorities Institute <a href="https://philarchive.org/rec/THOERP">working paper</a> - Existential risk pessimism and the time of perils</li><li>Ethics <a href="https://www.journals.uchicago.edu/doi/abs/10.1086/731436?journalCode=et">article</a> - Mistakes in the Moral Mathematics of Existential Risk</li><li>Philosophy &amp; Public Affairs <a href="https://onlinelibrary.wiley.com/doi/10.1111/papa.12248">article</a> - High Risk, Low Reward: A Challenge to the Astronomical Value of Existential Risk Mitigation</li><li>Toby Ord <a href="https://theprecipice.com/">Book</a> - The Precipice</li><li>Rethink Priorities <a href="https://rethinkpriorities.org/research-area/charting-the-precipice-the-time-of-perils-and-prioritizing-x-risk/">blogpost</a> - Charting the precipice</li><li>AI Futures Project <a href="https://ai-2027.com/">blogpost</a> - AI 2027</li></ul><p><strong>Trump's Higher Education </strong><strong>Threat</strong><strong> Compact</strong></p><ul><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Compact_for_Academic_Excellence_in_Higher_Education">article</a> - Compact for Academic Excellence in Higher Education</li><li>Pen America <a href="https://pen.org/trumps-compact-for-higher-education-faq/">explainer</a> - What is Trump’s Compact for Higher Education? And More Frequently Asked Questions</li><li><a href="https://vanderbiltaaup.org/2025/10/06/statement-by-the-vanderbilt-aaup-executive-committee-on-the-compact-for-academic-excellence-in-higher-education/">Statement</a> by the Vanderbilt AAUP Executive Committee on the “Compact for Academic Excellence in Higher Education”</li><li>The Vanderbilt Hustler <a href="https://vanderbilthustler.com/2025/10/20/breaking-chancellor-daniel-diermeier-fails-to-reject-higher-education-compact-reaffirms-vanderbilts-values-and-openness-to-discussion/">article</a> - BREAKING: Chancellor Daniel Diermeier fails to reject higher education compact, reaffirms Vanderbilt’s values and openness to discussion</li><li>The Vanderbilt Hustler <a href="https://vanderbilthustler.com/2025/11/06/students-and-faculty-organize-rally-outside-kirkland-hall-against-trump-administrations-higher-education-compact/">article</a> - Students and faculty organize rally outside Kirkland Hall against Trump administration’s higher education compact</li><li>Free Speech Center <a href="https://firstamendment.mtsu.edu/article/compact-for-academic-excellence/">article</a> - Compact for Academic Excellence</li></ul><p><strong>More of David's Work</strong></p><ul><li>Global Priorities Institute <a href="https://www.globalprioritiesinstitute.org/wp-content/uploads/David-Thorstad-What-power-seeking-theorems-do-not-show.pdf">working paper</a> - What power-seeking theorems do not show</li><li><a href="https://global.oup.com/academic/product/essays-on-longtermism-9780192883858?cc=us&amp;lang=en&amp;">Book</a> - Essays on Longtermism</li></ul><p><strong>Vibe Shift</strong></p><ul><li>Blood in the Machine <a href="https://www.bloodinthemachine.com/p/gpt-5-is-a-joke-will-it-matter">article</a> - GPT-5 Is a Joke. Will It Matter?</li><li>Futurism <a href="https://futurism.com/gpt-5-underwhelming">article</a> - Evidence Grows That GPT-5 Is a Bit of a Dud</li><li>Gary Marcus <a href="https://garymarcus.substack.com/p/gpt-5-overdue-overhyped-and-underwhelming">substack</a> - GPT-5: Overdue, overhyped and underwhelming. And that’s not the worst of it.</li><li>Pew Research <a href="https://www.pewresearch.org/internet/2025/04/03/how-the-us-public-and-ai-experts-view-artificial-intelligence/">report</a> - How the U.S. Public and AI Experts View Artificial Intelligence</li><li>N...</li></ul>]]>
      </itunes:summary>
      <itunes:keywords>AI, AI safety, technology, the singularity, singularity, artificial intelligence, effective altruism, philosophy, longtermism, superintelligence, AI skepticism, AI hype, AI funding, AI safety funding, philanthropy, open philanthropy, coefficient giving, global priorities institute, vanderbilt</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Guest" href="https://www.dthorstad.com" img="https://img.transistorcdn.com/jEJRxA7tqymAcKH3mg5yVzsrOkyhbTnQDL0wiLhkzX0/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS84MmVl/ODY2ZDMyYmMyYmU1/Y2UzZGMzZWY2MjAw/ZWZhNC5wbmc.jpg">Dr. David Thorstad</podcast:person>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/5c0f0e45/transcript.txt" type="text/plain"/>
      <podcast:chapters url="https://share.transistor.fm/s/5c0f0e45/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Getting Agentic w/ Alistair Lowe-Norris</title>
      <itunes:episode>24</itunes:episode>
      <podcast:episode>24</podcast:episode>
      <itunes:title>Getting Agentic w/ Alistair Lowe-Norris</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">6a0a2091-0747-4deb-bad3-48ab59fa4fbe</guid>
      <link>https://kairos.fm/intoaisafety/e024</link>
      <description>
        <![CDATA[<p>Alistair Lowe-Norris, Chief Responsible AI Officer at Iridius and co-host of The Agentic Insider podcast, joins to discuss AI compliance standards, the importance of narrowly scoping systems, and how procurement requirements could encourage responsible AI adoption across industries. We explore the gap between the empty promises companies provide and actual safety practices, as well as the importance of vigilance and continuous oversight.</p><p>Listen to Alistair on his <a href="https://agenticinsider.show/">podcast</a>, The Agentic Insider!</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><strong>Chapters<br></strong></p><ul><li>(00:00) - Intro</li>
<li>(02:46) - Trustworthy AI and the Human Side of Change</li>
<li>(13:57) - This is Essentially Avatar, Right?</li>
<li>(23:00) - AI Call Centers</li>
<li>(49:38) - Standards, Audits, and Accountability</li>
<li>(01:04:11) - What Happens when Standards aren’t Met?</li>
</ul><br><strong>Links</strong><ul><li>Iridius <a href="https://iridius.ai/">website</a></li></ul><p><strong>GPT-5 Commentary</strong></p><ul><li>Where's Your Ed At <a href="https://www.wheresyoured.at/how-does-gpt-5-work/">blogpost</a> - How Does GPT-5 Work?</li><li>Zvi LessWrong <a href="https://www.lesswrong.com/posts/eFd7NZ4KpYLM4ocBv/gpt-5-the-reverse-deepseek-moment">blogpost</a> - GPT-5: The Reverse DeepSeek moment</li><li>Blood in the Machine <a href="https://www.bloodinthemachine.com/p/gpt-5-is-a-joke-will-it-matter">article</a> - GPT-5 Is a Joke. Will It Matter?</li><li>Futurism <a href="https://futurism.com/gpt-5-underwhelming">article</a> - Evidence Grows That GPT-5 Is a Bit of a Dud</li><li>Gary Marcus <a href="https://garymarcus.substack.com/p/gpt-5-overdue-overhyped-and-underwhelming">substack</a> - GPT-5: Overdue, overhyped and underwhelming. And that’s not the worst of it.</li></ul><p><strong>Customer Service and AI Adoption</strong></p><ul><li>Gartner <a href="https://www.gartner.com/en/newsroom/press-releases/2024-07-09-gartner-survey-finds-64-percent-of-customers-would-prefer-that-companies-didnt-use-ai-for-customer-service">press release</a> - Gartner Survey Finds 64% of Customers Would Prefer That Companies Didn't Use AI for Customer Service</li><li><a href="https://arxiv.org/abs/2504.06145">Preprint</a> - Deploying Chatbots in Customer Service: Adoption Hurdles and Simple Remedies</li><li>KDD '25 <a href="https://dl.acm.org/doi/10.1145/3711896.3736557">paper</a> - Retrieval And Structuring Augmented Generation with Large Language Models</li><li>Global Nerdy <a href="https://www.globalnerdy.com/2024/04/16/retrieval-augmented-generation-explained-star-wars-style/">blogpost</a> - Retrieval-augmented generation explained “Star Wars” style</li><li>The Security Cafe <a href="https://securitycafe.io/p/quick-dirty-guide-soc2-compliance">article</a> - A Quick And Dirty Guide To Starting SOC2</li></ul><p><strong>Standards</strong></p><ul><li>ISO <a href="https://www.iso.org/artificial-intelligence/ai-management-systems">overview</a> - AI management systems</li><li>ISO <a href="https://www.iso.org/standard/42001">standard</a> - ISO/IEC 42001<ul><li>CyberZoni <a href="https://cyberzoni.com/standards/iso-42001/">guide</a> - ISO 42001 The Complete Guide</li><li>A-LIGN <a href="https://www.a-lign.com/articles/understanding-iso-42001">article</a> - Understanding ISO 42001</li></ul></li><li>ISO <a href="https://www.iso.org/standard/27001">standard</a> - ISO/IEC 27001</li><li>ISO <a href="https://www.iso.org/standard/42005">standard</a> - ISO/IEC 42005</li></ul><p><strong>Governance and Regulation</strong></p><ul><li>NIST <a href="https://www.nist.gov/itl/ai-risk-management-framework">framework</a> - AI Risk Management Framework</li><li>EU AI Act <a href="https://artificialintelligenceact.eu/article/99/">article</a> - Article 99: Penalties</li><li>Colorado Senate Bill 24-205 (Colorado AI Act) <a href="https://leg.colorado.gov/bills/sb24-205">webpage</a></li><li>Utah Senate Bill 149 <a href="https://le.utah.gov/~2024/bills/static/SB0149.html">webpage</a></li></ul><p><strong>Microsoft AI Compliance</strong></p><ul><li>Schellman <a href="https://www.schellman.com/blog/privacy/microsoft-dpr-ai-requirements-and-iso-42001">blogpost</a> - Microsoft DPR AI Requirements and ISO 42001</li><li>Microsoft <a href="https://learn.microsoft.com/en-us/compliance/regulatory/offering-iso-42001">documentation</a> - ISO/IEC 42001 AI Management System offering</li><li>Microsoft <a href="https://www.microsoft.com/ai/principles-and-approach">webpage</a> - Responsible AI Principles and Approach</li><li>Microsoft Service Trust Portal <a href="https://servicetrust.microsoft.com/DocumentPage/227a27b7-30c5-4110-8c8b-eb57a113e10f">documentation</a> - Responsible AI Standard v2</li><li>Microsoft <a href="https://cdn-dynmedia-1.microsoft.com/is/content/microsoftcorp/microsoft/accex/documents/presentations/FY25-Program-Guide-v11_en-US.pdf">documentation</a> - Supplier Security &amp; Privacy Assurance Program Guide v11 April 2025</li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Alistair Lowe-Norris, Chief Responsible AI Officer at Iridius and co-host of The Agentic Insider podcast, joins to discuss AI compliance standards, the importance of narrowly scoping systems, and how procurement requirements could encourage responsible AI adoption across industries. We explore the gap between the empty promises companies provide and actual safety practices, as well as the importance of vigilance and continuous oversight.</p><p>Listen to Alistair on his <a href="https://agenticinsider.show/">podcast</a>, The Agentic Insider!</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><strong>Chapters<br></strong></p><ul><li>(00:00) - Intro</li>
<li>(02:46) - Trustworthy AI and the Human Side of Change</li>
<li>(13:57) - This is Essentially Avatar, Right?</li>
<li>(23:00) - AI Call Centers</li>
<li>(49:38) - Standards, Audits, and Accountability</li>
<li>(01:04:11) - What Happens when Standards aren’t Met?</li>
</ul><br><strong>Links</strong><ul><li>Iridius <a href="https://iridius.ai/">website</a></li></ul><p><strong>GPT-5 Commentary</strong></p><ul><li>Where's Your Ed At <a href="https://www.wheresyoured.at/how-does-gpt-5-work/">blogpost</a> - How Does GPT-5 Work?</li><li>Zvi LessWrong <a href="https://www.lesswrong.com/posts/eFd7NZ4KpYLM4ocBv/gpt-5-the-reverse-deepseek-moment">blogpost</a> - GPT-5: The Reverse DeepSeek moment</li><li>Blood in the Machine <a href="https://www.bloodinthemachine.com/p/gpt-5-is-a-joke-will-it-matter">article</a> - GPT-5 Is a Joke. Will It Matter?</li><li>Futurism <a href="https://futurism.com/gpt-5-underwhelming">article</a> - Evidence Grows That GPT-5 Is a Bit of a Dud</li><li>Gary Marcus <a href="https://garymarcus.substack.com/p/gpt-5-overdue-overhyped-and-underwhelming">substack</a> - GPT-5: Overdue, overhyped and underwhelming. And that’s not the worst of it.</li></ul><p><strong>Customer Service and AI Adoption</strong></p><ul><li>Gartner <a href="https://www.gartner.com/en/newsroom/press-releases/2024-07-09-gartner-survey-finds-64-percent-of-customers-would-prefer-that-companies-didnt-use-ai-for-customer-service">press release</a> - Gartner Survey Finds 64% of Customers Would Prefer That Companies Didn't Use AI for Customer Service</li><li><a href="https://arxiv.org/abs/2504.06145">Preprint</a> - Deploying Chatbots in Customer Service: Adoption Hurdles and Simple Remedies</li><li>KDD '25 <a href="https://dl.acm.org/doi/10.1145/3711896.3736557">paper</a> - Retrieval And Structuring Augmented Generation with Large Language Models</li><li>Global Nerdy <a href="https://www.globalnerdy.com/2024/04/16/retrieval-augmented-generation-explained-star-wars-style/">blogpost</a> - Retrieval-augmented generation explained “Star Wars” style</li><li>The Security Cafe <a href="https://securitycafe.io/p/quick-dirty-guide-soc2-compliance">article</a> - A Quick And Dirty Guide To Starting SOC2</li></ul><p><strong>Standards</strong></p><ul><li>ISO <a href="https://www.iso.org/artificial-intelligence/ai-management-systems">overview</a> - AI management systems</li><li>ISO <a href="https://www.iso.org/standard/42001">standard</a> - ISO/IEC 42001<ul><li>CyberZoni <a href="https://cyberzoni.com/standards/iso-42001/">guide</a> - ISO 42001 The Complete Guide</li><li>A-LIGN <a href="https://www.a-lign.com/articles/understanding-iso-42001">article</a> - Understanding ISO 42001</li></ul></li><li>ISO <a href="https://www.iso.org/standard/27001">standard</a> - ISO/IEC 27001</li><li>ISO <a href="https://www.iso.org/standard/42005">standard</a> - ISO/IEC 42005</li></ul><p><strong>Governance and Regulation</strong></p><ul><li>NIST <a href="https://www.nist.gov/itl/ai-risk-management-framework">framework</a> - AI Risk Management Framework</li><li>EU AI Act <a href="https://artificialintelligenceact.eu/article/99/">article</a> - Article 99: Penalties</li><li>Colorado Senate Bill 24-205 (Colorado AI Act) <a href="https://leg.colorado.gov/bills/sb24-205">webpage</a></li><li>Utah Senate Bill 149 <a href="https://le.utah.gov/~2024/bills/static/SB0149.html">webpage</a></li></ul><p><strong>Microsoft AI Compliance</strong></p><ul><li>Schellman <a href="https://www.schellman.com/blog/privacy/microsoft-dpr-ai-requirements-and-iso-42001">blogpost</a> - Microsoft DPR AI Requirements and ISO 42001</li><li>Microsoft <a href="https://learn.microsoft.com/en-us/compliance/regulatory/offering-iso-42001">documentation</a> - ISO/IEC 42001 AI Management System offering</li><li>Microsoft <a href="https://www.microsoft.com/ai/principles-and-approach">webpage</a> - Responsible AI Principles and Approach</li><li>Microsoft Service Trust Portal <a href="https://servicetrust.microsoft.com/DocumentPage/227a27b7-30c5-4110-8c8b-eb57a113e10f">documentation</a> - Responsible AI Standard v2</li><li>Microsoft <a href="https://cdn-dynmedia-1.microsoft.com/is/content/microsoftcorp/microsoft/accex/documents/presentations/FY25-Program-Guide-v11_en-US.pdf">documentation</a> - Supplier Security &amp; Privacy Assurance Program Guide v11 April 2025</li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 20 Oct 2025 14:30:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/32031ccd/a274f471.mp3" length="68661997" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/ikaTBFhiK0QGougCg6ttHhgv6H7gj_Vcuq534nU7Pt4/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS83MmY5/OWEyMzI4YmI1Zjc1/NDhkMjZiZDA5Mzgx/MWQ5MS5wbmc.jpg"/>
      <itunes:duration>4290</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Alistair Lowe-Norris, Chief Responsible AI Officer at Iridius and co-host of The Agentic Insider podcast, joins to discuss AI compliance standards, the importance of narrowly scoping systems, and how procurement requirements could encourage responsible AI adoption across industries. We explore the gap between the empty promises companies provide and actual safety practices, as well as the importance of vigilance and continuous oversight.</p><p>Listen to Alistair on his <a href="https://agenticinsider.show/">podcast</a>, The Agentic Insider!</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><strong>Chapters<br></strong></p><ul><li>(00:00) - Intro</li>
<li>(02:46) - Trustworthy AI and the Human Side of Change</li>
<li>(13:57) - This is Essentially Avatar, Right?</li>
<li>(23:00) - AI Call Centers</li>
<li>(49:38) - Standards, Audits, and Accountability</li>
<li>(01:04:11) - What Happens when Standards aren’t Met?</li>
</ul><br><strong>Links</strong><ul><li>Iridius <a href="https://iridius.ai/">website</a></li></ul><p><strong>GPT-5 Commentary</strong></p><ul><li>Where's Your Ed At <a href="https://www.wheresyoured.at/how-does-gpt-5-work/">blogpost</a> - How Does GPT-5 Work?</li><li>Zvi LessWrong <a href="https://www.lesswrong.com/posts/eFd7NZ4KpYLM4ocBv/gpt-5-the-reverse-deepseek-moment">blogpost</a> - GPT-5: The Reverse DeepSeek moment</li><li>Blood in the Machine <a href="https://www.bloodinthemachine.com/p/gpt-5-is-a-joke-will-it-matter">article</a> - GPT-5 Is a Joke. Will It Matter?</li><li>Futurism <a href="https://futurism.com/gpt-5-underwhelming">article</a> - Evidence Grows That GPT-5 Is a Bit of a Dud</li><li>Gary Marcus <a href="https://garymarcus.substack.com/p/gpt-5-overdue-overhyped-and-underwhelming">substack</a> - GPT-5: Overdue, overhyped and underwhelming. And that’s not the worst of it.</li></ul><p><strong>Customer Service and AI Adoption</strong></p><ul><li>Gartner <a href="https://www.gartner.com/en/newsroom/press-releases/2024-07-09-gartner-survey-finds-64-percent-of-customers-would-prefer-that-companies-didnt-use-ai-for-customer-service">press release</a> - Gartner Survey Finds 64% of Customers Would Prefer That Companies Didn't Use AI for Customer Service</li><li><a href="https://arxiv.org/abs/2504.06145">Preprint</a> - Deploying Chatbots in Customer Service: Adoption Hurdles and Simple Remedies</li><li>KDD '25 <a href="https://dl.acm.org/doi/10.1145/3711896.3736557">paper</a> - Retrieval And Structuring Augmented Generation with Large Language Models</li><li>Global Nerdy <a href="https://www.globalnerdy.com/2024/04/16/retrieval-augmented-generation-explained-star-wars-style/">blogpost</a> - Retrieval-augmented generation explained “Star Wars” style</li><li>The Security Cafe <a href="https://securitycafe.io/p/quick-dirty-guide-soc2-compliance">article</a> - A Quick And Dirty Guide To Starting SOC2</li></ul><p><strong>Standards</strong></p><ul><li>ISO <a href="https://www.iso.org/artificial-intelligence/ai-management-systems">overview</a> - AI management systems</li><li>ISO <a href="https://www.iso.org/standard/42001">standard</a> - ISO/IEC 42001<ul><li>CyberZoni <a href="https://cyberzoni.com/standards/iso-42001/">guide</a> - ISO 42001 The Complete Guide</li><li>A-LIGN <a href="https://www.a-lign.com/articles/understanding-iso-42001">article</a> - Understanding ISO 42001</li></ul></li><li>ISO <a href="https://www.iso.org/standard/27001">standard</a> - ISO/IEC 27001</li><li>ISO <a href="https://www.iso.org/standard/42005">standard</a> - ISO/IEC 42005</li></ul><p><strong>Governance and Regulation</strong></p><ul><li>NIST <a href="https://www.nist.gov/itl/ai-risk-management-framework">framework</a> - AI Risk Management Framework</li><li>EU AI Act <a href="https://artificialintelligenceact.eu/article/99/">article</a> - Article 99: Penalties</li><li>Colorado Senate Bill 24-205 (Colorado AI Act) <a href="https://leg.colorado.gov/bills/sb24-205">webpage</a></li><li>Utah Senate Bill 149 <a href="https://le.utah.gov/~2024/bills/static/SB0149.html">webpage</a></li></ul><p><strong>Microsoft AI Compliance</strong></p><ul><li>Schellman <a href="https://www.schellman.com/blog/privacy/microsoft-dpr-ai-requirements-and-iso-42001">blogpost</a> - Microsoft DPR AI Requirements and ISO 42001</li><li>Microsoft <a href="https://learn.microsoft.com/en-us/compliance/regulatory/offering-iso-42001">documentation</a> - ISO/IEC 42001 AI Management System offering</li><li>Microsoft <a href="https://www.microsoft.com/ai/principles-and-approach">webpage</a> - Responsible AI Principles and Approach</li><li>Microsoft Service Trust Portal <a href="https://servicetrust.microsoft.com/DocumentPage/227a27b7-30c5-4110-8c8b-eb57a113e10f">documentation</a> - Responsible AI Standard v2</li><li>Microsoft <a href="https://cdn-dynmedia-1.microsoft.com/is/content/microsoftcorp/microsoft/accex/documents/presentations/FY25-Program-Guide-v11_en-US.pdf">documentation</a> - Supplier Security &amp; Privacy Assurance Program Guide v11 April 2025</li></ul>]]>
      </itunes:summary>
      <itunes:keywords>AI, AI safety, AI governance, AI compliance, Agentic AI, ISO 42001, ISO, Responsible AI, AI leadership, Alistair Lowe Norris, Alistair Lowe-Norris, compliance, auditing, AI auditing, AI compliance, AI standards, AI regulation, standards</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://intoaisafety.transistor.fm/people/alistair-lowe-norris" img="https://img.transistorcdn.com/zKrIto11gGokqeI8sfm3oEGzcmv2q7N1pS_QAyKOKIA/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9jYTY2/Yjc0NGNiYjcyMThl/Mzk2NjNlOThhYzQ0/NDYzOC5qcGVn.jpg">Alistair Lowe-Norris</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/32031ccd/transcript.json" type="application/json"/>
      <podcast:transcript url="https://share.transistor.fm/s/32031ccd/transcript.vtt" type="text/vtt" rel="captions"/>
      <podcast:transcript url="https://share.transistor.fm/s/32031ccd/transcript.txt" type="text/plain"/>
      <podcast:chapters url="https://share.transistor.fm/s/32031ccd/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Growing BlueDot's Impact w/ Li-Lian Ang</title>
      <itunes:episode>23</itunes:episode>
      <podcast:episode>23</podcast:episode>
      <itunes:title>Growing BlueDot's Impact w/ Li-Lian Ang</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">32b8731b-d8de-472e-b02f-284a3d86505c</guid>
      <link>https://kairos.fm/intoaisafety/e023/</link>
      <description>
        <![CDATA[<p>I'm joined by my good friend, Li-Lian Ang, first hire and product manager at BlueDot Impact. We discuss how BlueDot has evolved from their original course offerings to a new "defense-in-depth" approach, which focuses on three core threat models: reduced oversight in high risk scenarios (e.g. accelerated warfare), catastrophic terrorism (e.g. rogue actors with bioweapons), and the concentration of wealth and power (e.g. supercharged surveillance states). On top of that, we cover how BlueDot's strategies account for and reduce the negative impacts of common issues in AI safety, including exclusionary tendencies, elitism, and echo chambers.</p><p><strong><em>2025.09.15: </em></strong><em>Learn more about how to make design effective interventions to make AI go well and potentially even get funded for it on BlueDot Impact's </em><a href="https://bluedot.org/courses/agi-strategy"><em>AGI Strategy course</em></a><em>! BlueDot is also </em><a href="https://bluedot.org/join-us"><em>hiring</em></a><em>, so if you think you’d be a good fit, I definitely recommend applying; I had a great experience when I contracted as a course facilitator. If you do end up applying, let them know you found out about the opportunity from the podcast!</em></p><p><br>Follow Li-Lian on <a href="https://www.linkedin.com/in/anglilian/">LinkedIn</a>, and look at more of her work on her <a href="https://anglilian.com/">blog</a>!</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><br></p><ul><li>(03:23) - Meeting Through the Course</li>
<li>(05:46) - Eating Your Own Dog Food</li>
<li>(13:13) - Impact Acceleration</li>
<li>(22:13) - Breaking Out of the AI Safety Mold</li>
<li>(26:06) - Bluedot’s Risk Framework</li>
<li>(41:38) - Dangers of "Frontier" Models</li>
<li>(54:06) - The Need for AI Safety Advocates</li>
<li>(01:00:11) - Hot Takes and Pet Peeves</li>
</ul><br><strong>Links</strong><ul><li>BlueDot Impact <a href="https://bluedot.org/">website</a></li></ul><p><strong>Defense-in-Depth</strong></p><ul><li>BlueDot Impact <a href="https://bluedot.org/blog/course-portfolio-vision">blogpost</a> - Our vision for comprehensive AI safety training</li><li>Engineering for Humans <a href="https://www.engineeringforhumans.com/systems-engineering/the-swiss-cheese-model-designing-to-reduce-catastrophic-losses/">blogpost</a> - The Swiss cheese model: Designing to reduce catastrophic losses</li><li>Open Journal of Safety Science and Technology <a href="https://www.scirp.org/journal/paperinformation?paperid=70457">article</a> - The Evolution of Defense in Depth Approach: A Cross Sectorial Analysis</li></ul><p><strong>X-clusion and X-risk</strong></p><ul><li>Nature <a href="https://arxiv.org/abs/2502.09288">article</a> - AI Safety for Everyone</li><li>Ben Kuhn <a href="https://www.benkuhn.net/welcoming/">blogpost</a> - On being welcoming</li><li>Reflective Altruism <a href="https://reflectivealtruism.com/2023/01/12/off-series-that-bostrom-email/">blogpost</a> - Belonging (Part 1: That Bostrom email)</li></ul><p><strong>AIxBio</strong></p><ul><li>RAND <a href="https://www.rand.org/pubs/research_reports/RRA2977-2.html">report</a> - The Operational Risks of AI in Large-Scale Biological Attacks</li><li>OpenAI <a href="https://openai.com/index/building-an-early-warning-system-for-llm-aided-biological-threat-creation/">"publication"</a> (press release) - Building an early warning system for LLM-aided biological threat creation</li><li>Anthropic Frontier AI Red Team <a href="https://red.anthropic.com/2025/biorisk/">blogpost</a> - Why do we take LLMs seriously as a potential source of biorisk?</li><li>Kevin Esvelt <a href="https://arxiv.org/pdf/2503.15182">preprint</a> - Foundation models may exhibit staged progression in novel CBRN threat disclosure</li><li>Anthropic <a href="https://www.anthropic.com/news/activating-asl3-protections">press release</a> - Activating AI Safety Level 3 protections</li></ul><p><strong>Persuasive AI</strong></p><ul><li><a href="https://arxiv.org/abs/2412.17128v1">Preprint</a> - Lies, Damned Lies, and Distributional Language Statistics: Persuasion and Deception with Large Language Models</li><li>Nature Human Behavior <a href="https://www.nature.com/articles/s41562-025-02194-6?error=cookies_not_supported&amp;code=32f522a9-c964-4efa-9565-e013fbef4d63">article</a> - On the conversational persuasiveness of GPT-4</li><li><a href="https://arxiv.org/abs/2505.09662">Preprint</a> - Large Language Models Are More Persuasive Than Incentivized Human Persuaders</li></ul><p><strong>AI, Anthropomorphization, and Mental Health</strong></p><ul><li>Western News <a href="https://news.westernu.ca/2025/08/danger-of-anthropomorphic-ai/">article</a> - Expert insight: Humanlike chatbots detract from developing AI for the human good</li><li>AI &amp; Society <a href="https://link.springer.com/article/10.1007/s00146-022-01492-1">article</a> - Anthropomorphization and beyond: conceptualizing humanwashing of AI-enabled machines</li><li>Artificial Ignorance <a href="https://www.ignorance.ai/p/the-chatbot-trap">article</a> - The Chatbot Trap</li><li>Making Noise and Hearing Things <a href="https://makingnoiseandhearingthings.com/2022/08/03/large-language-models-cannot-replace-mental-health-professionals/">blogpost</a> - Large language models cannot replace mental health professionals</li><li>Idealogo <a href="https://www.brightfama.com/blog/2025/08/28/4-reasons-not-to-turn-chatgpt-into-your-therapist/">blogpost</a> - 4 reasons not to turn ChatGPT into your therapist</li><li>Journal of Medical Society <a href="https://journals.lww.com/jmso/fulltext/2024/38010/importance_of_informed_consent_in_medical_practice.1.aspx">Editorial</a> - Importance of informed consent in medical practice</li><li>Indian Journal of Medical Research <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC7055160/">article</a> - Consent in psychiatry - concept, application &amp; implications</li><li>Media Naama <a href="https://www.medianama.com/2025/04/223-chatgpt-sycophantic-tone-risks-humanizing-ai-chatbots/">article</a> - The Risk of Humanising AI Chabots: Why ChatGPT Mimicking Feelings Can Backfire</li><li>Becker's Behavioral Health <a href="https://www.beckersbehavioralhealth.com/ai-2/openais-mental-health-roadmap-5-things-to-know/">blogpost</a> - OpenAI’s mental health roadmap: 5 things to know</li></ul><p><strong>Miscellaneous References</strong></p><ul><li>Carnegie Council <a href="https://carnegiecouncil.org/media/article/what-do-we-mean-when-we-talk-about-ai-democratization">blogpost</a> - What Do We Mean When We Talk About "AI Democratization"?</li><li>Collective Intelligence Project <a href="https://www.cip.org/research/democratizing-ai">policy brief</a> - Four Approaches to Democratizing AI</li><li>BlueDot Impact <a href="https://bluedot.org/blog/how-does-ai-learn">blogpost</a> - How Does AI Learn? A Beginner's Guide with Examples</li><li>BlueDot Impact <a href="https://bluedot.org/blog/ai-safety-advocacy">blogpost</a> - AI safety needs more public-facing advocacy</li></ul><p><strong>More Li-Lian Links</strong></p><ul><li>Humans of Minerva podcast <a href="https://humansofminerva.com/">website</a></li><li>Li-Lian's <a href="https://www.goodreads.com/book/show/57027582-purple-is-the-noblest-shroud">book</a> - Purple is the Noblest Shroud</li></ul><p><strong>Relevant Podcasts from Kairos.fm</strong></p><ul><li><a href="https://kairos.fm/intoaisafety/e019/">Scaling Democracy w/ Dr. Igor Krawczuk</a> for AI safety exclusion and echo chambers</li><li><a href="https://kairos.fm/intoaisafety/e021/">Getting into PauseAI w/ Will Petillo</a> for AI in warfare and exclusion in AI safety</li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>I'm joined by my good friend, Li-Lian Ang, first hire and product manager at BlueDot Impact. We discuss how BlueDot has evolved from their original course offerings to a new "defense-in-depth" approach, which focuses on three core threat models: reduced oversight in high risk scenarios (e.g. accelerated warfare), catastrophic terrorism (e.g. rogue actors with bioweapons), and the concentration of wealth and power (e.g. supercharged surveillance states). On top of that, we cover how BlueDot's strategies account for and reduce the negative impacts of common issues in AI safety, including exclusionary tendencies, elitism, and echo chambers.</p><p><strong><em>2025.09.15: </em></strong><em>Learn more about how to make design effective interventions to make AI go well and potentially even get funded for it on BlueDot Impact's </em><a href="https://bluedot.org/courses/agi-strategy"><em>AGI Strategy course</em></a><em>! BlueDot is also </em><a href="https://bluedot.org/join-us"><em>hiring</em></a><em>, so if you think you’d be a good fit, I definitely recommend applying; I had a great experience when I contracted as a course facilitator. If you do end up applying, let them know you found out about the opportunity from the podcast!</em></p><p><br>Follow Li-Lian on <a href="https://www.linkedin.com/in/anglilian/">LinkedIn</a>, and look at more of her work on her <a href="https://anglilian.com/">blog</a>!</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><br></p><ul><li>(03:23) - Meeting Through the Course</li>
<li>(05:46) - Eating Your Own Dog Food</li>
<li>(13:13) - Impact Acceleration</li>
<li>(22:13) - Breaking Out of the AI Safety Mold</li>
<li>(26:06) - Bluedot’s Risk Framework</li>
<li>(41:38) - Dangers of "Frontier" Models</li>
<li>(54:06) - The Need for AI Safety Advocates</li>
<li>(01:00:11) - Hot Takes and Pet Peeves</li>
</ul><br><strong>Links</strong><ul><li>BlueDot Impact <a href="https://bluedot.org/">website</a></li></ul><p><strong>Defense-in-Depth</strong></p><ul><li>BlueDot Impact <a href="https://bluedot.org/blog/course-portfolio-vision">blogpost</a> - Our vision for comprehensive AI safety training</li><li>Engineering for Humans <a href="https://www.engineeringforhumans.com/systems-engineering/the-swiss-cheese-model-designing-to-reduce-catastrophic-losses/">blogpost</a> - The Swiss cheese model: Designing to reduce catastrophic losses</li><li>Open Journal of Safety Science and Technology <a href="https://www.scirp.org/journal/paperinformation?paperid=70457">article</a> - The Evolution of Defense in Depth Approach: A Cross Sectorial Analysis</li></ul><p><strong>X-clusion and X-risk</strong></p><ul><li>Nature <a href="https://arxiv.org/abs/2502.09288">article</a> - AI Safety for Everyone</li><li>Ben Kuhn <a href="https://www.benkuhn.net/welcoming/">blogpost</a> - On being welcoming</li><li>Reflective Altruism <a href="https://reflectivealtruism.com/2023/01/12/off-series-that-bostrom-email/">blogpost</a> - Belonging (Part 1: That Bostrom email)</li></ul><p><strong>AIxBio</strong></p><ul><li>RAND <a href="https://www.rand.org/pubs/research_reports/RRA2977-2.html">report</a> - The Operational Risks of AI in Large-Scale Biological Attacks</li><li>OpenAI <a href="https://openai.com/index/building-an-early-warning-system-for-llm-aided-biological-threat-creation/">"publication"</a> (press release) - Building an early warning system for LLM-aided biological threat creation</li><li>Anthropic Frontier AI Red Team <a href="https://red.anthropic.com/2025/biorisk/">blogpost</a> - Why do we take LLMs seriously as a potential source of biorisk?</li><li>Kevin Esvelt <a href="https://arxiv.org/pdf/2503.15182">preprint</a> - Foundation models may exhibit staged progression in novel CBRN threat disclosure</li><li>Anthropic <a href="https://www.anthropic.com/news/activating-asl3-protections">press release</a> - Activating AI Safety Level 3 protections</li></ul><p><strong>Persuasive AI</strong></p><ul><li><a href="https://arxiv.org/abs/2412.17128v1">Preprint</a> - Lies, Damned Lies, and Distributional Language Statistics: Persuasion and Deception with Large Language Models</li><li>Nature Human Behavior <a href="https://www.nature.com/articles/s41562-025-02194-6?error=cookies_not_supported&amp;code=32f522a9-c964-4efa-9565-e013fbef4d63">article</a> - On the conversational persuasiveness of GPT-4</li><li><a href="https://arxiv.org/abs/2505.09662">Preprint</a> - Large Language Models Are More Persuasive Than Incentivized Human Persuaders</li></ul><p><strong>AI, Anthropomorphization, and Mental Health</strong></p><ul><li>Western News <a href="https://news.westernu.ca/2025/08/danger-of-anthropomorphic-ai/">article</a> - Expert insight: Humanlike chatbots detract from developing AI for the human good</li><li>AI &amp; Society <a href="https://link.springer.com/article/10.1007/s00146-022-01492-1">article</a> - Anthropomorphization and beyond: conceptualizing humanwashing of AI-enabled machines</li><li>Artificial Ignorance <a href="https://www.ignorance.ai/p/the-chatbot-trap">article</a> - The Chatbot Trap</li><li>Making Noise and Hearing Things <a href="https://makingnoiseandhearingthings.com/2022/08/03/large-language-models-cannot-replace-mental-health-professionals/">blogpost</a> - Large language models cannot replace mental health professionals</li><li>Idealogo <a href="https://www.brightfama.com/blog/2025/08/28/4-reasons-not-to-turn-chatgpt-into-your-therapist/">blogpost</a> - 4 reasons not to turn ChatGPT into your therapist</li><li>Journal of Medical Society <a href="https://journals.lww.com/jmso/fulltext/2024/38010/importance_of_informed_consent_in_medical_practice.1.aspx">Editorial</a> - Importance of informed consent in medical practice</li><li>Indian Journal of Medical Research <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC7055160/">article</a> - Consent in psychiatry - concept, application &amp; implications</li><li>Media Naama <a href="https://www.medianama.com/2025/04/223-chatgpt-sycophantic-tone-risks-humanizing-ai-chatbots/">article</a> - The Risk of Humanising AI Chabots: Why ChatGPT Mimicking Feelings Can Backfire</li><li>Becker's Behavioral Health <a href="https://www.beckersbehavioralhealth.com/ai-2/openais-mental-health-roadmap-5-things-to-know/">blogpost</a> - OpenAI’s mental health roadmap: 5 things to know</li></ul><p><strong>Miscellaneous References</strong></p><ul><li>Carnegie Council <a href="https://carnegiecouncil.org/media/article/what-do-we-mean-when-we-talk-about-ai-democratization">blogpost</a> - What Do We Mean When We Talk About "AI Democratization"?</li><li>Collective Intelligence Project <a href="https://www.cip.org/research/democratizing-ai">policy brief</a> - Four Approaches to Democratizing AI</li><li>BlueDot Impact <a href="https://bluedot.org/blog/how-does-ai-learn">blogpost</a> - How Does AI Learn? A Beginner's Guide with Examples</li><li>BlueDot Impact <a href="https://bluedot.org/blog/ai-safety-advocacy">blogpost</a> - AI safety needs more public-facing advocacy</li></ul><p><strong>More Li-Lian Links</strong></p><ul><li>Humans of Minerva podcast <a href="https://humansofminerva.com/">website</a></li><li>Li-Lian's <a href="https://www.goodreads.com/book/show/57027582-purple-is-the-noblest-shroud">book</a> - Purple is the Noblest Shroud</li></ul><p><strong>Relevant Podcasts from Kairos.fm</strong></p><ul><li><a href="https://kairos.fm/intoaisafety/e019/">Scaling Democracy w/ Dr. Igor Krawczuk</a> for AI safety exclusion and echo chambers</li><li><a href="https://kairos.fm/intoaisafety/e021/">Getting into PauseAI w/ Will Petillo</a> for AI in warfare and exclusion in AI safety</li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 15 Sep 2025 09:14:55 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/693171f2/df040908.mp3" length="65004117" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/RHGvhylPBXvzREw1UP5mKmPpeaTtg6mSK91hCldXCNw/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS83MDNi/NDAzMWVkODI3Mzg2/NmJmYjU0OTViNmIx/OGJjMS5wbmc.jpg"/>
      <itunes:duration>4061</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>I'm joined by my good friend, Li-Lian Ang, first hire and product manager at BlueDot Impact. We discuss how BlueDot has evolved from their original course offerings to a new "defense-in-depth" approach, which focuses on three core threat models: reduced oversight in high risk scenarios (e.g. accelerated warfare), catastrophic terrorism (e.g. rogue actors with bioweapons), and the concentration of wealth and power (e.g. supercharged surveillance states). On top of that, we cover how BlueDot's strategies account for and reduce the negative impacts of common issues in AI safety, including exclusionary tendencies, elitism, and echo chambers.</p><p><strong><em>2025.09.15: </em></strong><em>Learn more about how to make design effective interventions to make AI go well and potentially even get funded for it on BlueDot Impact's </em><a href="https://bluedot.org/courses/agi-strategy"><em>AGI Strategy course</em></a><em>! BlueDot is also </em><a href="https://bluedot.org/join-us"><em>hiring</em></a><em>, so if you think you’d be a good fit, I definitely recommend applying; I had a great experience when I contracted as a course facilitator. If you do end up applying, let them know you found out about the opportunity from the podcast!</em></p><p><br>Follow Li-Lian on <a href="https://www.linkedin.com/in/anglilian/">LinkedIn</a>, and look at more of her work on her <a href="https://anglilian.com/">blog</a>!</p><p>As part of my effort to make this whole podcasting thing more sustainable, I have created a Kairos.fm <a href="https://www.patreon.com/cw/Kairosfm">Patreon</a> which includes an extended version of this episode. Supporting gets you access to these extended cuts, as well as other perks in development.</p><p><br></p><ul><li>(03:23) - Meeting Through the Course</li>
<li>(05:46) - Eating Your Own Dog Food</li>
<li>(13:13) - Impact Acceleration</li>
<li>(22:13) - Breaking Out of the AI Safety Mold</li>
<li>(26:06) - Bluedot’s Risk Framework</li>
<li>(41:38) - Dangers of "Frontier" Models</li>
<li>(54:06) - The Need for AI Safety Advocates</li>
<li>(01:00:11) - Hot Takes and Pet Peeves</li>
</ul><br><strong>Links</strong><ul><li>BlueDot Impact <a href="https://bluedot.org/">website</a></li></ul><p><strong>Defense-in-Depth</strong></p><ul><li>BlueDot Impact <a href="https://bluedot.org/blog/course-portfolio-vision">blogpost</a> - Our vision for comprehensive AI safety training</li><li>Engineering for Humans <a href="https://www.engineeringforhumans.com/systems-engineering/the-swiss-cheese-model-designing-to-reduce-catastrophic-losses/">blogpost</a> - The Swiss cheese model: Designing to reduce catastrophic losses</li><li>Open Journal of Safety Science and Technology <a href="https://www.scirp.org/journal/paperinformation?paperid=70457">article</a> - The Evolution of Defense in Depth Approach: A Cross Sectorial Analysis</li></ul><p><strong>X-clusion and X-risk</strong></p><ul><li>Nature <a href="https://arxiv.org/abs/2502.09288">article</a> - AI Safety for Everyone</li><li>Ben Kuhn <a href="https://www.benkuhn.net/welcoming/">blogpost</a> - On being welcoming</li><li>Reflective Altruism <a href="https://reflectivealtruism.com/2023/01/12/off-series-that-bostrom-email/">blogpost</a> - Belonging (Part 1: That Bostrom email)</li></ul><p><strong>AIxBio</strong></p><ul><li>RAND <a href="https://www.rand.org/pubs/research_reports/RRA2977-2.html">report</a> - The Operational Risks of AI in Large-Scale Biological Attacks</li><li>OpenAI <a href="https://openai.com/index/building-an-early-warning-system-for-llm-aided-biological-threat-creation/">"publication"</a> (press release) - Building an early warning system for LLM-aided biological threat creation</li><li>Anthropic Frontier AI Red Team <a href="https://red.anthropic.com/2025/biorisk/">blogpost</a> - Why do we take LLMs seriously as a potential source of biorisk?</li><li>Kevin Esvelt <a href="https://arxiv.org/pdf/2503.15182">preprint</a> - Foundation models may exhibit staged progression in novel CBRN threat disclosure</li><li>Anthropic <a href="https://www.anthropic.com/news/activating-asl3-protections">press release</a> - Activating AI Safety Level 3 protections</li></ul><p><strong>Persuasive AI</strong></p><ul><li><a href="https://arxiv.org/abs/2412.17128v1">Preprint</a> - Lies, Damned Lies, and Distributional Language Statistics: Persuasion and Deception with Large Language Models</li><li>Nature Human Behavior <a href="https://www.nature.com/articles/s41562-025-02194-6?error=cookies_not_supported&amp;code=32f522a9-c964-4efa-9565-e013fbef4d63">article</a> - On the conversational persuasiveness of GPT-4</li><li><a href="https://arxiv.org/abs/2505.09662">Preprint</a> - Large Language Models Are More Persuasive Than Incentivized Human Persuaders</li></ul><p><strong>AI, Anthropomorphization, and Mental Health</strong></p><ul><li>Western News <a href="https://news.westernu.ca/2025/08/danger-of-anthropomorphic-ai/">article</a> - Expert insight: Humanlike chatbots detract from developing AI for the human good</li><li>AI &amp; Society <a href="https://link.springer.com/article/10.1007/s00146-022-01492-1">article</a> - Anthropomorphization and beyond: conceptualizing humanwashing of AI-enabled machines</li><li>Artificial Ignorance <a href="https://www.ignorance.ai/p/the-chatbot-trap">article</a> - The Chatbot Trap</li><li>Making Noise and Hearing Things <a href="https://makingnoiseandhearingthings.com/2022/08/03/large-language-models-cannot-replace-mental-health-professionals/">blogpost</a> - Large language models cannot replace mental health professionals</li><li>Idealogo <a href="https://www.brightfama.com/blog/2025/08/28/4-reasons-not-to-turn-chatgpt-into-your-therapist/">blogpost</a> - 4 reasons not to turn ChatGPT into your therapist</li><li>Journal of Medical Society <a href="https://journals.lww.com/jmso/fulltext/2024/38010/importance_of_informed_consent_in_medical_practice.1.aspx">Editorial</a> - Importance of informed consent in medical practice</li><li>Indian Journal of Medical Research <a href="https://pmc.ncbi.nlm.nih.gov/articles/PMC7055160/">article</a> - Consent in psychiatry - concept, application &amp; implications</li><li>Media Naama <a href="https://www.medianama.com/2025/04/223-chatgpt-sycophantic-tone-risks-humanizing-ai-chatbots/">article</a> - The Risk of Humanising AI Chabots: Why ChatGPT Mimicking Feelings Can Backfire</li><li>Becker's Behavioral Health <a href="https://www.beckersbehavioralhealth.com/ai-2/openais-mental-health-roadmap-5-things-to-know/">blogpost</a> - OpenAI’s mental health roadmap: 5 things to know</li></ul><p><strong>Miscellaneous References</strong></p><ul><li>Carnegie Council <a href="https://carnegiecouncil.org/media/article/what-do-we-mean-when-we-talk-about-ai-democratization">blogpost</a> - What Do We Mean When We Talk About "AI Democratization"?</li><li>Collective Intelligence Project <a href="https://www.cip.org/research/democratizing-ai">policy brief</a> - Four Approaches to Democratizing AI</li><li>BlueDot Impact <a href="https://bluedot.org/blog/how-does-ai-learn">blogpost</a> - How Does AI Learn? A Beginner's Guide with Examples</li><li>BlueDot Impact <a href="https://bluedot.org/blog/ai-safety-advocacy">blogpost</a> - AI safety needs more public-facing advocacy</li></ul><p><strong>More Li-Lian Links</strong></p><ul><li>Humans of Minerva podcast <a href="https://humansofminerva.com/">website</a></li><li>Li-Lian's <a href="https://www.goodreads.com/book/show/57027582-purple-is-the-noblest-shroud">book</a> - Purple is the Noblest Shroud</li></ul><p><strong>Relevant Podcasts from Kairos.fm</strong></p><ul><li><a href="https://kairos.fm/intoaisafety/e019/">Scaling Democracy w/ Dr. Igor Krawczuk</a> for AI safety exclusion and echo chambers</li><li><a href="https://kairos.fm/intoaisafety/e021/">Getting into PauseAI w/ Will Petillo</a> for AI in warfare and exclusion in AI safety</li></ul>]]>
      </itunes:summary>
      <itunes:keywords>AI, artificial intelligence, AI safety, responsible AI, AI safety training, BlueDot Impact, BlueDot, Intro to Transformative AI, AI safety strategy, AIxBio, Defense in depth, Defense-in-depth, swiss cheese model, swiss cheese method</itunes:keywords>
      <itunes:explicit>Yes</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://anglilian.com" img="https://img.transistorcdn.com/PjLFAQTxrPr45IM_76IxuwugPO7paG296-o83kLTyg8/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8wY2Jh/MDM2MTRjOGU0ZjNm/Mzg3MmE1NmViNThm/ZDFhNi5qcGVn.jpg">Li-Lian Ang</podcast:person>
      <podcast:chapters url="https://share.transistor.fm/s/693171f2/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Layoffs to Leadership w/ Andres Sepulveda Morales</title>
      <itunes:episode>22</itunes:episode>
      <podcast:episode>22</podcast:episode>
      <itunes:title>Layoffs to Leadership w/ Andres Sepulveda Morales</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">a7322f60-bd7f-4726-be49-6bf9c8fa54b9</guid>
      <link>https://kairos.fm/intoaisafety/e022</link>
      <description>
        <![CDATA[<p>Andres Sepulveda Morales joins me to discuss his journey from three tech layoffs to founding Red Mage Creative and leading the Fort Collins chapter of the Rocky Mountain AI Interest Group (RMAIIG). We explore the current tech job market, AI anxiety in nonprofits, dark patterns in AI systems, and building inclusive tech communities that welcome diverse perspectives.</p><p><br>Reach out to Andres on his <a href="https://www.linkedin.com/in/andres-sepulveda-morales/">LinkedIn</a>, or check out the Red Mage Creative <a href="https://redmage.cc/">website</a>!</p><p>For any listeners in Colorado, consider attending an RMAIIG event: <a href="https://www.meetup.com/rmaiig/">Boulder</a>; <a href="https://www.meetup.com/fort-collins-ai-for-everyone-rmaiig/events/307983383/?eventOrigin=group_upcoming_events">Fort Collins<br></a><br></p><ul><li>(00:00) - Intro</li>
<li>(01:04) - Andres' Journey</li>
<li>(05:15) - Tech Layoff Cycle</li>
<li>(26:12) - Why AI?</li>
<li>(30:58) - What is Red Mage?</li>
<li>(36:12) - AI as a Tool</li>
<li>(41:55) - AInxiety</li>
<li>(47:26) - Dark Patterns and Critical Perspectives</li>
<li>(01:01:35) - RMAIIG</li>
<li>(01:10:09) - Inclusive Tech Education</li>
<li>(01:18:05) - Colorado AI Governance</li>
<li>(01:23:46) - Building Your Own Tech Community</li>
</ul><br><strong>Links</strong><p><strong>Tech Job Market</strong></p><ul><li>Layoff tracker <a href="https://layoffs.fyi/">website</a></li><li>The Big Newsletter <a href="https://www.thebignewsletter.com/p/why-are-we-pretending-ai-is-going">article</a> - Why Are We Pretending AI Is Going to Take All the Jobs?</li><li>METR <a href="https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/">preprint</a> - Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity</li><li>AI Business <a href="https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis">blogpost</a> - <a href="https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis">https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis</a></li><li>Crunchbase <a href="https://news.crunchbase.com/layoffs/big-tech-leads-workforce-cuts-msft-amzn/">article</a> - Data: Tech Layoffs Remain Stubbornly High, With Big Tech Leading The Way</li><li>Computerworld <a href="https://www.computerworld.com/article/4003261/tech-layoffs-surge-even-as-us-unemployment-remains-stable.html">article</a> - Tech layoffs surge even as US unemployment remains stable</li><li>Apollo Technical <a href="https://www.apollotechnical.com/ghost-jobs-in-tech-why-companies-are-posting-roles-they-dont-plan-to-fill/">blogpost</a> - Ghost jobs in tech: Why companies are posting roles they don’t plan to fill</li><li>The HR Digest <a href="https://www.thehrdigest.com/the-rise-of-ghost-jobs-is-leaving-job-seekers-frustrated-and-disappointed/">article</a> - The Rise of Ghost Jobs Is Leaving Job Seekers Frustrated and Disappointed</li><li>A Life After Layoff <a href="https://www.youtube.com/watch?v=rp06jQIYQ-w">video</a> - The Tech Job Market Is Hot Trash Right Now</li><li>Economy Media <a href="https://www.youtube.com/watch?v=ppYtD18CokI">video</a> - Will The Tech Job Market Ever Recover?</li><li>Soleyman Shahir <a href="https://www.youtube.com/watch?v=X0aYO8GMB4c">video</a> - Tech CEO Explains: The Real Reason Behind AI Layoffs</li></ul><p><strong>Dark Patterns</strong></p><ul><li>Deceptive Design <a href="https://www.deceptive.design/">website</a></li><li>Journal of Legal Analysis <a href="https://www.researchgate.net/publication/350340175_Shining_a_Light_on_Dark_Patterns">article</a> - Shining a Light on Dark Patterns</li><li>ICLR <a href="https://arxiv.org/abs/2503.10728">paper</a> - DarkBench: Benchmarking Dark Patterns in Large Language Models</li><li>Computing Within Limits <a href="https://computingwithinlimits.org/2025/papers/limits2025-beigon-imposing-ai.pdf">paper</a> - Imposing AI: Deceptive design patterns against sustainability</li><li>Communications of the ACM <a href="https://cacm.acm.org/practice/dark-patterns/">blogpost</a> - Dark Patterns</li><li>[Preprint] - A Comprehensive Study on Dark Patterns</li></ul><p><strong>Colorado AI Regulation</strong></p><ul><li>Senate Bill 24-205 (Colorado AI Act) <a href="https://leg.colorado.gov/sites/default/files/2024a_205_signed.pdf">bill</a> and <a href="https://leg.colorado.gov/bills/sb24-205">webpage</a></li><li>NAAG <a href="https://www.naag.org/attorney-general-journal/a-deep-dive-into-colorados-artificial-intelligence-act/">article</a> - A Deep Dive into Colorado’s Artificial Intelligence Act</li><li>Colorado Sun <a href="https://coloradosun.com/2025/04/23/colorado-artificial-intelligence-law-ai/">article</a> - Why Colorado’s artificial intelligence law is a big deal for the whole country</li><li>CFO Dive <a href="https://www.cfodive.com/news/a-heavy-lift-colorado-ai-law-sets-high-bar-analysts-say/753025/">blogpost</a> - ‘Heavy lift’: Colorado AI law sets high bar, analysts say</li><li>Denver 7 <a href="https://www.denver7.com/money/science-and-tech/colorado-could-lose-funding-as-trump-administration-targets-ai-regulations">article</a> - Colorado could lose federal funding as Trump administration targets AI regulations</li><li>America's AI Action Plan <a href="https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdf">document</a></li></ul><p><strong>Other Sources</strong></p><ul><li>Concordia Framework <a href="https://arxiv.org/abs/2312.03664">report</a> and <a href="https://github.com/google-deepmind/concordia">repo</a></li><li>80,000 Hours <a href="https://80000hours.org/">website</a></li><li>AI Incident Database <a href="https://incidentdatabase.ai/">website</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Andres Sepulveda Morales joins me to discuss his journey from three tech layoffs to founding Red Mage Creative and leading the Fort Collins chapter of the Rocky Mountain AI Interest Group (RMAIIG). We explore the current tech job market, AI anxiety in nonprofits, dark patterns in AI systems, and building inclusive tech communities that welcome diverse perspectives.</p><p><br>Reach out to Andres on his <a href="https://www.linkedin.com/in/andres-sepulveda-morales/">LinkedIn</a>, or check out the Red Mage Creative <a href="https://redmage.cc/">website</a>!</p><p>For any listeners in Colorado, consider attending an RMAIIG event: <a href="https://www.meetup.com/rmaiig/">Boulder</a>; <a href="https://www.meetup.com/fort-collins-ai-for-everyone-rmaiig/events/307983383/?eventOrigin=group_upcoming_events">Fort Collins<br></a><br></p><ul><li>(00:00) - Intro</li>
<li>(01:04) - Andres' Journey</li>
<li>(05:15) - Tech Layoff Cycle</li>
<li>(26:12) - Why AI?</li>
<li>(30:58) - What is Red Mage?</li>
<li>(36:12) - AI as a Tool</li>
<li>(41:55) - AInxiety</li>
<li>(47:26) - Dark Patterns and Critical Perspectives</li>
<li>(01:01:35) - RMAIIG</li>
<li>(01:10:09) - Inclusive Tech Education</li>
<li>(01:18:05) - Colorado AI Governance</li>
<li>(01:23:46) - Building Your Own Tech Community</li>
</ul><br><strong>Links</strong><p><strong>Tech Job Market</strong></p><ul><li>Layoff tracker <a href="https://layoffs.fyi/">website</a></li><li>The Big Newsletter <a href="https://www.thebignewsletter.com/p/why-are-we-pretending-ai-is-going">article</a> - Why Are We Pretending AI Is Going to Take All the Jobs?</li><li>METR <a href="https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/">preprint</a> - Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity</li><li>AI Business <a href="https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis">blogpost</a> - <a href="https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis">https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis</a></li><li>Crunchbase <a href="https://news.crunchbase.com/layoffs/big-tech-leads-workforce-cuts-msft-amzn/">article</a> - Data: Tech Layoffs Remain Stubbornly High, With Big Tech Leading The Way</li><li>Computerworld <a href="https://www.computerworld.com/article/4003261/tech-layoffs-surge-even-as-us-unemployment-remains-stable.html">article</a> - Tech layoffs surge even as US unemployment remains stable</li><li>Apollo Technical <a href="https://www.apollotechnical.com/ghost-jobs-in-tech-why-companies-are-posting-roles-they-dont-plan-to-fill/">blogpost</a> - Ghost jobs in tech: Why companies are posting roles they don’t plan to fill</li><li>The HR Digest <a href="https://www.thehrdigest.com/the-rise-of-ghost-jobs-is-leaving-job-seekers-frustrated-and-disappointed/">article</a> - The Rise of Ghost Jobs Is Leaving Job Seekers Frustrated and Disappointed</li><li>A Life After Layoff <a href="https://www.youtube.com/watch?v=rp06jQIYQ-w">video</a> - The Tech Job Market Is Hot Trash Right Now</li><li>Economy Media <a href="https://www.youtube.com/watch?v=ppYtD18CokI">video</a> - Will The Tech Job Market Ever Recover?</li><li>Soleyman Shahir <a href="https://www.youtube.com/watch?v=X0aYO8GMB4c">video</a> - Tech CEO Explains: The Real Reason Behind AI Layoffs</li></ul><p><strong>Dark Patterns</strong></p><ul><li>Deceptive Design <a href="https://www.deceptive.design/">website</a></li><li>Journal of Legal Analysis <a href="https://www.researchgate.net/publication/350340175_Shining_a_Light_on_Dark_Patterns">article</a> - Shining a Light on Dark Patterns</li><li>ICLR <a href="https://arxiv.org/abs/2503.10728">paper</a> - DarkBench: Benchmarking Dark Patterns in Large Language Models</li><li>Computing Within Limits <a href="https://computingwithinlimits.org/2025/papers/limits2025-beigon-imposing-ai.pdf">paper</a> - Imposing AI: Deceptive design patterns against sustainability</li><li>Communications of the ACM <a href="https://cacm.acm.org/practice/dark-patterns/">blogpost</a> - Dark Patterns</li><li>[Preprint] - A Comprehensive Study on Dark Patterns</li></ul><p><strong>Colorado AI Regulation</strong></p><ul><li>Senate Bill 24-205 (Colorado AI Act) <a href="https://leg.colorado.gov/sites/default/files/2024a_205_signed.pdf">bill</a> and <a href="https://leg.colorado.gov/bills/sb24-205">webpage</a></li><li>NAAG <a href="https://www.naag.org/attorney-general-journal/a-deep-dive-into-colorados-artificial-intelligence-act/">article</a> - A Deep Dive into Colorado’s Artificial Intelligence Act</li><li>Colorado Sun <a href="https://coloradosun.com/2025/04/23/colorado-artificial-intelligence-law-ai/">article</a> - Why Colorado’s artificial intelligence law is a big deal for the whole country</li><li>CFO Dive <a href="https://www.cfodive.com/news/a-heavy-lift-colorado-ai-law-sets-high-bar-analysts-say/753025/">blogpost</a> - ‘Heavy lift’: Colorado AI law sets high bar, analysts say</li><li>Denver 7 <a href="https://www.denver7.com/money/science-and-tech/colorado-could-lose-funding-as-trump-administration-targets-ai-regulations">article</a> - Colorado could lose federal funding as Trump administration targets AI regulations</li><li>America's AI Action Plan <a href="https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdf">document</a></li></ul><p><strong>Other Sources</strong></p><ul><li>Concordia Framework <a href="https://arxiv.org/abs/2312.03664">report</a> and <a href="https://github.com/google-deepmind/concordia">repo</a></li><li>80,000 Hours <a href="https://80000hours.org/">website</a></li><li>AI Incident Database <a href="https://incidentdatabase.ai/">website</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 04 Aug 2025 09:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/6dc56af2/dadec247.mp3" length="96003300" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/mInrd1yf6rnVm7fJUdLQBzi1L7_yvX3tskJtLC12NiQ/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS85MDll/M2E5NTcwMDA5OTMy/ZGRlNjZkNWI3M2Q1/NThlZS5wbmc.jpg"/>
      <itunes:duration>5999</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Andres Sepulveda Morales joins me to discuss his journey from three tech layoffs to founding Red Mage Creative and leading the Fort Collins chapter of the Rocky Mountain AI Interest Group (RMAIIG). We explore the current tech job market, AI anxiety in nonprofits, dark patterns in AI systems, and building inclusive tech communities that welcome diverse perspectives.</p><p><br>Reach out to Andres on his <a href="https://www.linkedin.com/in/andres-sepulveda-morales/">LinkedIn</a>, or check out the Red Mage Creative <a href="https://redmage.cc/">website</a>!</p><p>For any listeners in Colorado, consider attending an RMAIIG event: <a href="https://www.meetup.com/rmaiig/">Boulder</a>; <a href="https://www.meetup.com/fort-collins-ai-for-everyone-rmaiig/events/307983383/?eventOrigin=group_upcoming_events">Fort Collins<br></a><br></p><ul><li>(00:00) - Intro</li>
<li>(01:04) - Andres' Journey</li>
<li>(05:15) - Tech Layoff Cycle</li>
<li>(26:12) - Why AI?</li>
<li>(30:58) - What is Red Mage?</li>
<li>(36:12) - AI as a Tool</li>
<li>(41:55) - AInxiety</li>
<li>(47:26) - Dark Patterns and Critical Perspectives</li>
<li>(01:01:35) - RMAIIG</li>
<li>(01:10:09) - Inclusive Tech Education</li>
<li>(01:18:05) - Colorado AI Governance</li>
<li>(01:23:46) - Building Your Own Tech Community</li>
</ul><br><strong>Links</strong><p><strong>Tech Job Market</strong></p><ul><li>Layoff tracker <a href="https://layoffs.fyi/">website</a></li><li>The Big Newsletter <a href="https://www.thebignewsletter.com/p/why-are-we-pretending-ai-is-going">article</a> - Why Are We Pretending AI Is Going to Take All the Jobs?</li><li>METR <a href="https://metr.org/blog/2025-07-10-early-2025-ai-experienced-os-dev-study/">preprint</a> - Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity</li><li>AI Business <a href="https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis">blogpost</a> - <a href="https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis">https://aibusiness.com/responsible-ai/debunking-the-ai-job-crisis</a></li><li>Crunchbase <a href="https://news.crunchbase.com/layoffs/big-tech-leads-workforce-cuts-msft-amzn/">article</a> - Data: Tech Layoffs Remain Stubbornly High, With Big Tech Leading The Way</li><li>Computerworld <a href="https://www.computerworld.com/article/4003261/tech-layoffs-surge-even-as-us-unemployment-remains-stable.html">article</a> - Tech layoffs surge even as US unemployment remains stable</li><li>Apollo Technical <a href="https://www.apollotechnical.com/ghost-jobs-in-tech-why-companies-are-posting-roles-they-dont-plan-to-fill/">blogpost</a> - Ghost jobs in tech: Why companies are posting roles they don’t plan to fill</li><li>The HR Digest <a href="https://www.thehrdigest.com/the-rise-of-ghost-jobs-is-leaving-job-seekers-frustrated-and-disappointed/">article</a> - The Rise of Ghost Jobs Is Leaving Job Seekers Frustrated and Disappointed</li><li>A Life After Layoff <a href="https://www.youtube.com/watch?v=rp06jQIYQ-w">video</a> - The Tech Job Market Is Hot Trash Right Now</li><li>Economy Media <a href="https://www.youtube.com/watch?v=ppYtD18CokI">video</a> - Will The Tech Job Market Ever Recover?</li><li>Soleyman Shahir <a href="https://www.youtube.com/watch?v=X0aYO8GMB4c">video</a> - Tech CEO Explains: The Real Reason Behind AI Layoffs</li></ul><p><strong>Dark Patterns</strong></p><ul><li>Deceptive Design <a href="https://www.deceptive.design/">website</a></li><li>Journal of Legal Analysis <a href="https://www.researchgate.net/publication/350340175_Shining_a_Light_on_Dark_Patterns">article</a> - Shining a Light on Dark Patterns</li><li>ICLR <a href="https://arxiv.org/abs/2503.10728">paper</a> - DarkBench: Benchmarking Dark Patterns in Large Language Models</li><li>Computing Within Limits <a href="https://computingwithinlimits.org/2025/papers/limits2025-beigon-imposing-ai.pdf">paper</a> - Imposing AI: Deceptive design patterns against sustainability</li><li>Communications of the ACM <a href="https://cacm.acm.org/practice/dark-patterns/">blogpost</a> - Dark Patterns</li><li>[Preprint] - A Comprehensive Study on Dark Patterns</li></ul><p><strong>Colorado AI Regulation</strong></p><ul><li>Senate Bill 24-205 (Colorado AI Act) <a href="https://leg.colorado.gov/sites/default/files/2024a_205_signed.pdf">bill</a> and <a href="https://leg.colorado.gov/bills/sb24-205">webpage</a></li><li>NAAG <a href="https://www.naag.org/attorney-general-journal/a-deep-dive-into-colorados-artificial-intelligence-act/">article</a> - A Deep Dive into Colorado’s Artificial Intelligence Act</li><li>Colorado Sun <a href="https://coloradosun.com/2025/04/23/colorado-artificial-intelligence-law-ai/">article</a> - Why Colorado’s artificial intelligence law is a big deal for the whole country</li><li>CFO Dive <a href="https://www.cfodive.com/news/a-heavy-lift-colorado-ai-law-sets-high-bar-analysts-say/753025/">blogpost</a> - ‘Heavy lift’: Colorado AI law sets high bar, analysts say</li><li>Denver 7 <a href="https://www.denver7.com/money/science-and-tech/colorado-could-lose-funding-as-trump-administration-targets-ai-regulations">article</a> - Colorado could lose federal funding as Trump administration targets AI regulations</li><li>America's AI Action Plan <a href="https://www.whitehouse.gov/wp-content/uploads/2025/07/Americas-AI-Action-Plan.pdf">document</a></li></ul><p><strong>Other Sources</strong></p><ul><li>Concordia Framework <a href="https://arxiv.org/abs/2312.03664">report</a> and <a href="https://github.com/google-deepmind/concordia">repo</a></li><li>80,000 Hours <a href="https://80000hours.org/">website</a></li><li>AI Incident Database <a href="https://incidentdatabase.ai/">website</a></li></ul>]]>
      </itunes:summary>
      <itunes:keywords>AI, artificial intelligence, tech layoffs, dark patterns, dark patterns in AI, tech job market, Colorado AI, Colorado AI legislation, AI legislation, America's AI Action Plan, inclusivity, AI consulting, RMAIIG, Rocky Mountain AI Interest Group</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:chapters url="https://share.transistor.fm/s/6dc56af2/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Getting Into PauseAI w/ Will Petillo</title>
      <itunes:episode>21</itunes:episode>
      <podcast:episode>21</podcast:episode>
      <itunes:title>Getting Into PauseAI w/ Will Petillo</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">24daa207-d0a6-489f-9655-353db0daea99</guid>
      <link>https://kairos.fm/intoaisafety/e021/</link>
      <description>
        <![CDATA[<p>Will Petillo, onboarding team lead at <a href="https://pauseai.info/">PauseAI</a>, joins me to discuss the grassroots movement advocating for a pause on frontier AI model development. We explore PauseAI's strategy, talk about common misconceptions Will hears, and dig into how diverse perspectives still converge on the need to slow down AI development.</p><p><strong>Will's Links</strong></p><ul><li>Personal <a href="https://www.zenmarmotdigital.com/blog">blog</a> on AI</li><li>His <a href="https://www.figma.com/board/q1jTwOez837RG3SjGcfvlv/AI-Safety-Debate?node-id=0-1">mindmap</a> of the AI x-risk debate</li><li><a href="https://will9371.itch.io/">Game demos</a></li><li>AI focused YouTube <a href="https://youtube.com/@willpetillo1189?si=mt0z348YVus1JKfL">channel</a></li></ul><p><br></p><ul><li>(00:00) - Intro</li>
<li>(03:36) - What is PauseAI</li>
<li>(10:10) - Will Petillo's journey into AI safety advocacy</li>
<li>(21:13) - Understanding PauseAI</li>
<li>(31:35) - Pursuing a pause</li>
<li>(40:06) - Balancing advocacy in a complex world</li>
<li>(45:54) - Why a pause on frontier models?</li>
<li>(54:48) - Diverse perspectives within PauseAI</li>
<li>(59:55) - PauseAI misconceptions</li>
<li>(01:16:40) - Ongoing AI governance efforts (SB1047)</li>
<li>(01:28:52) - The role of incremental progress</li>
<li>(01:35:16) - Safety-washing and corporate responsibility</li>
<li>(01:37:23) - Lessons from environmentalism</li>
<li>(01:41:59) - Will's superlatives</li>
</ul><br><strong>Links</strong><ul><li><a href="https://pauseai.info/">PauseAI</a></li><li><a href="https://www.pauseai-us.org/">PauseAI-US</a></li></ul><p><strong>Related Kairos.fm Episodes</strong></p><ul><li>Into AI Safety <a href="https://kairos.fm/intoaisafety/e019/">episode</a> with Dr. Igor Krawczuk</li><li>muckrAIkers <a href="https://kairos.fm/muckraikers/e002/">episode</a> on SB1047</li></ul><p><strong>Exclusionary Tendencies</strong></p><ul><li>Jacobin <a href="https://jacobin.com/2023/01/effective-altruism-longtermism-nick-bostrom-racism">article</a> - Elite Universities Gave Us Effective Altruism, the Dumbest Idea of the Century</li><li>SSIR <a href="https://ssir.org/articles/entry/the_elitist_philanthropy_of_so_called_effective_altruism">article</a> - The Elitist Philanthropy of So-Called Effective Altruism</li><li>Persuasion <a href="https://www.persuasion.community/p/the-problem-with-effective-altruism">blogpost</a> - The Problem with Effective Altruism</li><li>Dark Markets <a href="https://davidzmorris.substack.com/p/whats-so-bad-about-rationalism">blogpost</a> - What's So Bad About Rationalism?</li><li>FEE <a href="https://fee.org/articles/whats-wrong-with-the-rationality-community/">blogpost</a> - What’s Wrong With the Rationality Community?</li></ul><p><strong>AI in Warfare</strong></p><ul><li>Master's <a href="https://apps.dtic.mil/sti/tr/pdf/ADA255221.pdf">Thesis</a> - The Evolution of Artificial Intelligence and Expert Computer Systems in the Army</li><li>International Journal of Intelligent Systems <a href="https://onlinelibrary.wiley.com/doi/10.1155/2023/8676366">article</a> - Artificial Intelligence in the Military: An Overview of the Capabilities, Applications, and Challenges</li><li><a href="https://arxiv.org/abs/2411.17840">Preprint</a> - Basic Research, Lethal Effects: Military AI Research Funding as Enlistment</li><li>AOAV <a href="https://aoav.org.uk/2019/military-age-males-in-us-drone-strikes/#:~:text=Military-age%20males%20refers%20to,boys%20and%20men%20with%20combatants">Article</a> - ‘Military Age Males’ in US Drone Strikes</li><li>The Conversation <a href="https://theconversation.com/gaza-war-israel-using-ai-to-identify-human-targets-raising-fears-that-innocents-are-being-caught-in-the-net-227422">article</a> - Gaza war: Israel using AI to identify human targets raising fears that innocents are being caught in the net</li><li>972 <a href="https://www.972mag.com/lavender-ai-israeli-army-gaza/">article</a> - ‘Lavender’: The AI machine directing Israel’s bombing spree in Gaza</li><li>IDF <a href="https://www.idf.il/210062">press release</a> - The IDF's Use of Data Technologies in Intelligence Processing</li><li>Lieber Institute West Point <a href="https://lieber.westpoint.edu/gospel-lavender-law-armed-conflict/">article</a> - Israel–Hamas 2024 Symposium</li><li>Verfassungsblog <a href="https://verfassungsblog.de/gaza-artificial-intelligence-and-kill-lists/">article</a> - Gaza, Artificial Intelligence, and Kill Lists</li><li>RAND <a href="https://www.rand.org/content/dam/rand/pubs/research_reports/RRA2600/RRA2679-1/RAND_RRA2679-1.pdf">research report</a> - Dr. Li Bicheng, or How China Learned to Stop Worrying and Love Social Media Manipulation</li><li>The Intercept <a href="https://theintercept.com/drone-papers/">article collection</a> - The Drone Papers</li><li>AFIT <a href="https://scholar.afit.edu/cgi/viewcontent.cgi?article=2572&amp;context=facpub">faculty publication</a> - On Large Language Models in National Security Applications</li><li>Nature <a href="https://link.springer.com/chapter/10.1057/978-1-137-55408-6_1?error=cookies_not_supported&amp;code=63211579-2434-43a1-bbad-974daff6cd7f">article</a> - Death by Metadata: The Bioinformationalisation of Life and the Transliteration of Algorithms to Flesh</li></ul><p><strong>Legislation</strong></p><ul><li>LegiScan <a href="https://legiscan.com/CA/text/SB1047/2023">page</a> on SB1047</li><li>NY State Senate <a href="https://www.nysenate.gov/legislation/bills/2025/A6453/amendment/A">page</a> on the RAISE Act</li><li>Congress <a href="https://www.congress.gov/bill/119th-congress/senate-bill/146">page</a> on the TAKE IT DOWN Act</li></ul><p><strong>The Gavernor</strong></p><ul><li>FastCompany <a href="https://www.fastcompany.com/91175604/big-tech-lobbyists-ai-safety-bill-gavin-newsom">article</a> - Big Tech may be focusing its lobbying push on the California AI safety bill’s last stop: Gavin Newsom</li><li>POLITICO <a href="https://www.politico.com/news/2024/10/01/newsom-silicon-valley-ai-safety-00181776">article</a> - How California politics killed a nationally important AI bill</li><li>Newsom's <a href="https://www.documentcloud.org/documents/25178071-sb-1047-veto-message/?responsive=1&amp;title=1">veto message</a></li><li>Additional relevant lobbying documentation - <a href="https://cal-access.sos.ca.gov/PDFGen/pdfgen.prg?filingid=2978661&amp;amendid=0">[1]</a>, <a href="https://cal-access.sos.ca.gov/PDFGen/pdfgen.prg?filingid=2935898&amp;amendid=1">[2]</a></li><li>Jacobin <a href="https://jacobin.com/2024/09/gavin-newsom-ai-tech-bill-sb-1047">article</a> - With Newsom’s Veto, Big Tech Beats Democracy</li></ul><p><strong>Misc. Links</strong></p><ul><li>FLI <a href="https://futureoflife.org/open-letter/pause-giant-ai-experiments/">Open Letter</a> on an AI pause</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Overton_window">article</a> - Overton window</li><li>Daniel Smachtenburger YouTube <a href="https://www.youtube.com/watch?v=4kBoLVvoqVY&amp;t=11s">video</a> - An Introduction to the Metacrisis</li><li>VAISU <a href="https://vaisu.ai/">website</a> (looks broken as of 2025.06.19)</li><li>AI Impacts <a href="https://aiimpacts.org/wp-content/uploads/2023/04/Why-Did-Environmentalism-Become-Partisan-1.pdf">report</a> - Why Did Environmentalism Become Partisan?</li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Will Petillo, onboarding team lead at <a href="https://pauseai.info/">PauseAI</a>, joins me to discuss the grassroots movement advocating for a pause on frontier AI model development. We explore PauseAI's strategy, talk about common misconceptions Will hears, and dig into how diverse perspectives still converge on the need to slow down AI development.</p><p><strong>Will's Links</strong></p><ul><li>Personal <a href="https://www.zenmarmotdigital.com/blog">blog</a> on AI</li><li>His <a href="https://www.figma.com/board/q1jTwOez837RG3SjGcfvlv/AI-Safety-Debate?node-id=0-1">mindmap</a> of the AI x-risk debate</li><li><a href="https://will9371.itch.io/">Game demos</a></li><li>AI focused YouTube <a href="https://youtube.com/@willpetillo1189?si=mt0z348YVus1JKfL">channel</a></li></ul><p><br></p><ul><li>(00:00) - Intro</li>
<li>(03:36) - What is PauseAI</li>
<li>(10:10) - Will Petillo's journey into AI safety advocacy</li>
<li>(21:13) - Understanding PauseAI</li>
<li>(31:35) - Pursuing a pause</li>
<li>(40:06) - Balancing advocacy in a complex world</li>
<li>(45:54) - Why a pause on frontier models?</li>
<li>(54:48) - Diverse perspectives within PauseAI</li>
<li>(59:55) - PauseAI misconceptions</li>
<li>(01:16:40) - Ongoing AI governance efforts (SB1047)</li>
<li>(01:28:52) - The role of incremental progress</li>
<li>(01:35:16) - Safety-washing and corporate responsibility</li>
<li>(01:37:23) - Lessons from environmentalism</li>
<li>(01:41:59) - Will's superlatives</li>
</ul><br><strong>Links</strong><ul><li><a href="https://pauseai.info/">PauseAI</a></li><li><a href="https://www.pauseai-us.org/">PauseAI-US</a></li></ul><p><strong>Related Kairos.fm Episodes</strong></p><ul><li>Into AI Safety <a href="https://kairos.fm/intoaisafety/e019/">episode</a> with Dr. Igor Krawczuk</li><li>muckrAIkers <a href="https://kairos.fm/muckraikers/e002/">episode</a> on SB1047</li></ul><p><strong>Exclusionary Tendencies</strong></p><ul><li>Jacobin <a href="https://jacobin.com/2023/01/effective-altruism-longtermism-nick-bostrom-racism">article</a> - Elite Universities Gave Us Effective Altruism, the Dumbest Idea of the Century</li><li>SSIR <a href="https://ssir.org/articles/entry/the_elitist_philanthropy_of_so_called_effective_altruism">article</a> - The Elitist Philanthropy of So-Called Effective Altruism</li><li>Persuasion <a href="https://www.persuasion.community/p/the-problem-with-effective-altruism">blogpost</a> - The Problem with Effective Altruism</li><li>Dark Markets <a href="https://davidzmorris.substack.com/p/whats-so-bad-about-rationalism">blogpost</a> - What's So Bad About Rationalism?</li><li>FEE <a href="https://fee.org/articles/whats-wrong-with-the-rationality-community/">blogpost</a> - What’s Wrong With the Rationality Community?</li></ul><p><strong>AI in Warfare</strong></p><ul><li>Master's <a href="https://apps.dtic.mil/sti/tr/pdf/ADA255221.pdf">Thesis</a> - The Evolution of Artificial Intelligence and Expert Computer Systems in the Army</li><li>International Journal of Intelligent Systems <a href="https://onlinelibrary.wiley.com/doi/10.1155/2023/8676366">article</a> - Artificial Intelligence in the Military: An Overview of the Capabilities, Applications, and Challenges</li><li><a href="https://arxiv.org/abs/2411.17840">Preprint</a> - Basic Research, Lethal Effects: Military AI Research Funding as Enlistment</li><li>AOAV <a href="https://aoav.org.uk/2019/military-age-males-in-us-drone-strikes/#:~:text=Military-age%20males%20refers%20to,boys%20and%20men%20with%20combatants">Article</a> - ‘Military Age Males’ in US Drone Strikes</li><li>The Conversation <a href="https://theconversation.com/gaza-war-israel-using-ai-to-identify-human-targets-raising-fears-that-innocents-are-being-caught-in-the-net-227422">article</a> - Gaza war: Israel using AI to identify human targets raising fears that innocents are being caught in the net</li><li>972 <a href="https://www.972mag.com/lavender-ai-israeli-army-gaza/">article</a> - ‘Lavender’: The AI machine directing Israel’s bombing spree in Gaza</li><li>IDF <a href="https://www.idf.il/210062">press release</a> - The IDF's Use of Data Technologies in Intelligence Processing</li><li>Lieber Institute West Point <a href="https://lieber.westpoint.edu/gospel-lavender-law-armed-conflict/">article</a> - Israel–Hamas 2024 Symposium</li><li>Verfassungsblog <a href="https://verfassungsblog.de/gaza-artificial-intelligence-and-kill-lists/">article</a> - Gaza, Artificial Intelligence, and Kill Lists</li><li>RAND <a href="https://www.rand.org/content/dam/rand/pubs/research_reports/RRA2600/RRA2679-1/RAND_RRA2679-1.pdf">research report</a> - Dr. Li Bicheng, or How China Learned to Stop Worrying and Love Social Media Manipulation</li><li>The Intercept <a href="https://theintercept.com/drone-papers/">article collection</a> - The Drone Papers</li><li>AFIT <a href="https://scholar.afit.edu/cgi/viewcontent.cgi?article=2572&amp;context=facpub">faculty publication</a> - On Large Language Models in National Security Applications</li><li>Nature <a href="https://link.springer.com/chapter/10.1057/978-1-137-55408-6_1?error=cookies_not_supported&amp;code=63211579-2434-43a1-bbad-974daff6cd7f">article</a> - Death by Metadata: The Bioinformationalisation of Life and the Transliteration of Algorithms to Flesh</li></ul><p><strong>Legislation</strong></p><ul><li>LegiScan <a href="https://legiscan.com/CA/text/SB1047/2023">page</a> on SB1047</li><li>NY State Senate <a href="https://www.nysenate.gov/legislation/bills/2025/A6453/amendment/A">page</a> on the RAISE Act</li><li>Congress <a href="https://www.congress.gov/bill/119th-congress/senate-bill/146">page</a> on the TAKE IT DOWN Act</li></ul><p><strong>The Gavernor</strong></p><ul><li>FastCompany <a href="https://www.fastcompany.com/91175604/big-tech-lobbyists-ai-safety-bill-gavin-newsom">article</a> - Big Tech may be focusing its lobbying push on the California AI safety bill’s last stop: Gavin Newsom</li><li>POLITICO <a href="https://www.politico.com/news/2024/10/01/newsom-silicon-valley-ai-safety-00181776">article</a> - How California politics killed a nationally important AI bill</li><li>Newsom's <a href="https://www.documentcloud.org/documents/25178071-sb-1047-veto-message/?responsive=1&amp;title=1">veto message</a></li><li>Additional relevant lobbying documentation - <a href="https://cal-access.sos.ca.gov/PDFGen/pdfgen.prg?filingid=2978661&amp;amendid=0">[1]</a>, <a href="https://cal-access.sos.ca.gov/PDFGen/pdfgen.prg?filingid=2935898&amp;amendid=1">[2]</a></li><li>Jacobin <a href="https://jacobin.com/2024/09/gavin-newsom-ai-tech-bill-sb-1047">article</a> - With Newsom’s Veto, Big Tech Beats Democracy</li></ul><p><strong>Misc. Links</strong></p><ul><li>FLI <a href="https://futureoflife.org/open-letter/pause-giant-ai-experiments/">Open Letter</a> on an AI pause</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Overton_window">article</a> - Overton window</li><li>Daniel Smachtenburger YouTube <a href="https://www.youtube.com/watch?v=4kBoLVvoqVY&amp;t=11s">video</a> - An Introduction to the Metacrisis</li><li>VAISU <a href="https://vaisu.ai/">website</a> (looks broken as of 2025.06.19)</li><li>AI Impacts <a href="https://aiimpacts.org/wp-content/uploads/2023/04/Why-Did-Environmentalism-Become-Partisan-1.pdf">report</a> - Why Did Environmentalism Become Partisan?</li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 23 Jun 2025 10:00:00 -0600</pubDate>
      <author>Jacob Haimes, Will Petillo</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/7ce784af/0e279ba9.mp3" length="78061070" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes, Will Petillo</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/II1LMWNhR37qiVibQNH8ETVNK2yBpVAMRoMuCibQ38c/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS81ZjZi/YmJlM2M4ZTkxNjBl/NmFiMTJmNDFlOWIx/YzU0Ni5wbmc.jpg"/>
      <itunes:duration>6484</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Will Petillo, onboarding team lead at <a href="https://pauseai.info/">PauseAI</a>, joins me to discuss the grassroots movement advocating for a pause on frontier AI model development. We explore PauseAI's strategy, talk about common misconceptions Will hears, and dig into how diverse perspectives still converge on the need to slow down AI development.</p><p><strong>Will's Links</strong></p><ul><li>Personal <a href="https://www.zenmarmotdigital.com/blog">blog</a> on AI</li><li>His <a href="https://www.figma.com/board/q1jTwOez837RG3SjGcfvlv/AI-Safety-Debate?node-id=0-1">mindmap</a> of the AI x-risk debate</li><li><a href="https://will9371.itch.io/">Game demos</a></li><li>AI focused YouTube <a href="https://youtube.com/@willpetillo1189?si=mt0z348YVus1JKfL">channel</a></li></ul><p><br></p><ul><li>(00:00) - Intro</li>
<li>(03:36) - What is PauseAI</li>
<li>(10:10) - Will Petillo's journey into AI safety advocacy</li>
<li>(21:13) - Understanding PauseAI</li>
<li>(31:35) - Pursuing a pause</li>
<li>(40:06) - Balancing advocacy in a complex world</li>
<li>(45:54) - Why a pause on frontier models?</li>
<li>(54:48) - Diverse perspectives within PauseAI</li>
<li>(59:55) - PauseAI misconceptions</li>
<li>(01:16:40) - Ongoing AI governance efforts (SB1047)</li>
<li>(01:28:52) - The role of incremental progress</li>
<li>(01:35:16) - Safety-washing and corporate responsibility</li>
<li>(01:37:23) - Lessons from environmentalism</li>
<li>(01:41:59) - Will's superlatives</li>
</ul><br><strong>Links</strong><ul><li><a href="https://pauseai.info/">PauseAI</a></li><li><a href="https://www.pauseai-us.org/">PauseAI-US</a></li></ul><p><strong>Related Kairos.fm Episodes</strong></p><ul><li>Into AI Safety <a href="https://kairos.fm/intoaisafety/e019/">episode</a> with Dr. Igor Krawczuk</li><li>muckrAIkers <a href="https://kairos.fm/muckraikers/e002/">episode</a> on SB1047</li></ul><p><strong>Exclusionary Tendencies</strong></p><ul><li>Jacobin <a href="https://jacobin.com/2023/01/effective-altruism-longtermism-nick-bostrom-racism">article</a> - Elite Universities Gave Us Effective Altruism, the Dumbest Idea of the Century</li><li>SSIR <a href="https://ssir.org/articles/entry/the_elitist_philanthropy_of_so_called_effective_altruism">article</a> - The Elitist Philanthropy of So-Called Effective Altruism</li><li>Persuasion <a href="https://www.persuasion.community/p/the-problem-with-effective-altruism">blogpost</a> - The Problem with Effective Altruism</li><li>Dark Markets <a href="https://davidzmorris.substack.com/p/whats-so-bad-about-rationalism">blogpost</a> - What's So Bad About Rationalism?</li><li>FEE <a href="https://fee.org/articles/whats-wrong-with-the-rationality-community/">blogpost</a> - What’s Wrong With the Rationality Community?</li></ul><p><strong>AI in Warfare</strong></p><ul><li>Master's <a href="https://apps.dtic.mil/sti/tr/pdf/ADA255221.pdf">Thesis</a> - The Evolution of Artificial Intelligence and Expert Computer Systems in the Army</li><li>International Journal of Intelligent Systems <a href="https://onlinelibrary.wiley.com/doi/10.1155/2023/8676366">article</a> - Artificial Intelligence in the Military: An Overview of the Capabilities, Applications, and Challenges</li><li><a href="https://arxiv.org/abs/2411.17840">Preprint</a> - Basic Research, Lethal Effects: Military AI Research Funding as Enlistment</li><li>AOAV <a href="https://aoav.org.uk/2019/military-age-males-in-us-drone-strikes/#:~:text=Military-age%20males%20refers%20to,boys%20and%20men%20with%20combatants">Article</a> - ‘Military Age Males’ in US Drone Strikes</li><li>The Conversation <a href="https://theconversation.com/gaza-war-israel-using-ai-to-identify-human-targets-raising-fears-that-innocents-are-being-caught-in-the-net-227422">article</a> - Gaza war: Israel using AI to identify human targets raising fears that innocents are being caught in the net</li><li>972 <a href="https://www.972mag.com/lavender-ai-israeli-army-gaza/">article</a> - ‘Lavender’: The AI machine directing Israel’s bombing spree in Gaza</li><li>IDF <a href="https://www.idf.il/210062">press release</a> - The IDF's Use of Data Technologies in Intelligence Processing</li><li>Lieber Institute West Point <a href="https://lieber.westpoint.edu/gospel-lavender-law-armed-conflict/">article</a> - Israel–Hamas 2024 Symposium</li><li>Verfassungsblog <a href="https://verfassungsblog.de/gaza-artificial-intelligence-and-kill-lists/">article</a> - Gaza, Artificial Intelligence, and Kill Lists</li><li>RAND <a href="https://www.rand.org/content/dam/rand/pubs/research_reports/RRA2600/RRA2679-1/RAND_RRA2679-1.pdf">research report</a> - Dr. Li Bicheng, or How China Learned to Stop Worrying and Love Social Media Manipulation</li><li>The Intercept <a href="https://theintercept.com/drone-papers/">article collection</a> - The Drone Papers</li><li>AFIT <a href="https://scholar.afit.edu/cgi/viewcontent.cgi?article=2572&amp;context=facpub">faculty publication</a> - On Large Language Models in National Security Applications</li><li>Nature <a href="https://link.springer.com/chapter/10.1057/978-1-137-55408-6_1?error=cookies_not_supported&amp;code=63211579-2434-43a1-bbad-974daff6cd7f">article</a> - Death by Metadata: The Bioinformationalisation of Life and the Transliteration of Algorithms to Flesh</li></ul><p><strong>Legislation</strong></p><ul><li>LegiScan <a href="https://legiscan.com/CA/text/SB1047/2023">page</a> on SB1047</li><li>NY State Senate <a href="https://www.nysenate.gov/legislation/bills/2025/A6453/amendment/A">page</a> on the RAISE Act</li><li>Congress <a href="https://www.congress.gov/bill/119th-congress/senate-bill/146">page</a> on the TAKE IT DOWN Act</li></ul><p><strong>The Gavernor</strong></p><ul><li>FastCompany <a href="https://www.fastcompany.com/91175604/big-tech-lobbyists-ai-safety-bill-gavin-newsom">article</a> - Big Tech may be focusing its lobbying push on the California AI safety bill’s last stop: Gavin Newsom</li><li>POLITICO <a href="https://www.politico.com/news/2024/10/01/newsom-silicon-valley-ai-safety-00181776">article</a> - How California politics killed a nationally important AI bill</li><li>Newsom's <a href="https://www.documentcloud.org/documents/25178071-sb-1047-veto-message/?responsive=1&amp;title=1">veto message</a></li><li>Additional relevant lobbying documentation - <a href="https://cal-access.sos.ca.gov/PDFGen/pdfgen.prg?filingid=2978661&amp;amendid=0">[1]</a>, <a href="https://cal-access.sos.ca.gov/PDFGen/pdfgen.prg?filingid=2935898&amp;amendid=1">[2]</a></li><li>Jacobin <a href="https://jacobin.com/2024/09/gavin-newsom-ai-tech-bill-sb-1047">article</a> - With Newsom’s Veto, Big Tech Beats Democracy</li></ul><p><strong>Misc. Links</strong></p><ul><li>FLI <a href="https://futureoflife.org/open-letter/pause-giant-ai-experiments/">Open Letter</a> on an AI pause</li><li>Wikipedia <a href="https://en.wikipedia.org/wiki/Overton_window">article</a> - Overton window</li><li>Daniel Smachtenburger YouTube <a href="https://www.youtube.com/watch?v=4kBoLVvoqVY&amp;t=11s">video</a> - An Introduction to the Metacrisis</li><li>VAISU <a href="https://vaisu.ai/">website</a> (looks broken as of 2025.06.19)</li><li>AI Impacts <a href="https://aiimpacts.org/wp-content/uploads/2023/04/Why-Did-Environmentalism-Become-Partisan-1.pdf">report</a> - Why Did Environmentalism Become Partisan?</li></ul>]]>
      </itunes:summary>
      <itunes:keywords>technology, AI, artificial intelligence, PauseAI, pause ai, AI pause, stop AI, AI risks</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/7ce784af/transcript.txt" type="text/plain"/>
      <podcast:chapters url="https://share.transistor.fm/s/7ce784af/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>Making Your Voice Heard w/ Tristan &amp; Felix de Simone</title>
      <itunes:episode>20</itunes:episode>
      <podcast:episode>20</podcast:episode>
      <itunes:title>Making Your Voice Heard w/ Tristan &amp; Felix de Simone</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">b93ca71c-29ca-4790-a3cb-bc885b8c2c76</guid>
      <link>https://kairos.fm/intoaisafety/e020/</link>
      <description>
        <![CDATA[<p>I am joined by Tristan Williams and Felix de Simone to discuss their work on the potential of constituent communication, specifically in the context of AI legislation. These two worked as part of an AI Safety Camp team to understand whether or not it would be useful for more people to be sharing their experiences, concerns, and opinions with their government representative (hint, it is).</p><p>Check out the <a href="https://forum.effectivealtruism.org/posts/5oStggnYLGzomhvvn/talking-to-congress-can-constituents-contacting-their">blogpost</a> on their findings, "Talking to Congress: Can constituents contacting their legislator influence policy?" and the <a href="https://www.guidedtrack.com/programs/e7xnz7q/run">tool</a> they created!</p><p></p><ul><li>(01:53) - Introductions</li>
<li>(04:04) - Starting the project</li>
<li>(13:30) - Project overview</li>
<li>(16:36) - Understanding constituent communication</li>
<li>(28:50) - Literature review</li>
<li>(35:52) - Phase 2</li>
<li>(43:26) - Creating a tool for citizen engagement</li>
<li>(50:16) - Crafting your message</li>
<li>(59:40) - The game of advocacy</li>
<li>(01:15:19) - Difficulties on the project</li>
<li>(01:22:33) - Call to action</li>
<li>(01:32:30) - Outro</li>
</ul><br><strong>Links</strong><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.aisafety.camp/">AI Safety Camp</a></li><li><a href="https://pauseai.info/">Pause AI</a></li><li><a href="https://bluedot.org/">BlueDot Impact</a></li><li>TIME <a href="https://time.com/6972134/ai-lobbying-tech-policy-surge/">article</a> - There’s an AI Lobbying Frenzy in Washington. Big Tech Is Dominating</li><li>Congressional Management Foundation <a href="https://www.congressfoundation.org/storage/documents/CMF_Pubs/cwc-perceptions-of-citizen-advocacy.pdf">study</a> - Communicating with Congress: Perceptions of Citizen Advocacy on Capitol Hill</li><li>Congressional Management Foundation <a href="https://www.congressfoundation.org/storage/documents/CMF_Pubs/cmf_citizen_engagement_rebuilding_democratic_dialogue.pdf">study</a> - The Future of Citizen Engagement: Rebuilding the Democratic Dialogue</li><li>Tristan and Felix's <a href="https://forum.effectivealtruism.org/posts/5oStggnYLGzomhvvn/talking-to-congress-can-constituents-contacting-their">blogpost</a> - Talking to Congress: Can constituents contacting their legislator influence policy?</li><li>Wired <a href="https://www.wired.com/story/opengov-report-congress-constituent-communication/">article</a> - What It Takes to Make Congress Actually Listen</li><li>American Journal of Polical Science <a href="https://projects.iq.harvard.edu/files/cces/files/AnsolabehereKuriwaki_AJPS.pdf">article</a> - Congressional Representation: Accountability from the Constituent’s Perspective</li><li>Political Behavior <a href="https://www.jstor.org/stable/43653416">article</a> - Call Your Legislator: A Field Experimental Study of the Impact of a Constituency Mobilization Campaign on Legislative Voting</li><li>Guided Track <a href="https://www.guidedtrack.com/">website</a></li><li>The <a href="https://www.guidedtrack.com/programs/e7xnz7q/run">Tool</a></li><li>Holistic AI global regulatory <a href="https://tracker.holisticai.com/feed">tracker</a></li><li>White &amp; Case global regulatory <a href="https://www.whitecase.com/insight-our-thinking/ai-watch-global-regulatory-tracker">tracker</a></li><li>Steptoe US AI legislation <a href="https://www.steptoe.com/en/ai-legislative-tracker.html">tracker</a></li><li>Manatt US AIxHealth legislation <a href="https://www.manatt.com/insights/newsletters/health-highlights/manatt-health-health-ai-policy-tracker">tracker</a></li><li>Issue One <a href="https://issueone.org/articles/big-tech-spent-record-sums-on-lobbying-last-year/">article</a> - Big Tech Cozies Up to New Administration After Spending Record Sums on Lobbying Last Year</li><li>Verfassungsblog <a href="https://verfassungsblog.de/bigtechs-efforts-to-derail-the-ai-act/">article</a> - BigTech’s Efforts to Derail the AI Act</li><li>MIT Technology Review <a href="https://www.technologyreview.com/2025/01/21/1110260/openai-ups-its-lobbying-efforts-nearly-seven-fold/">article</a> - OpenAI has upped its lobbying efforts nearly sevenfold</li><li>Open Secrets <a href="https://www.opensecrets.org/federal-lobbying/issues/lobbyists?id=SCI&amp;year=2023">webpage</a> - Issue Profile: Science &amp; Technology</li><li>Statista <a href="https://www.statista.com/statistics/257344/top-lobbying-spenders-in-the-us/">data</a> - Leading lobbying spenders in the United States in 2024</li><li>Global Justice Now <a href="https://www.globaljustice.org.uk/resource/democracy-at-risk-in-davos-new-report-exposes-big-tech-lobbying-and-political-interference/">report</a> - Democracy at risk in Davos: new report exposes big tech lobbying and political interference</li><li>Ipsos <a href="https://www.ipsos.com/en-us/where-americans-stand-ai">article</a> - Where Americans stand on AI</li><li>AP-NORC <a href="https://apnorc.org/projects/there-is-bipartisan-concern-about-the-use-of-ai-in-the-2024-elections/">report</a> - There Is Bipartisan Concern About the Use of AI in the 2024 Elections</li><li>AI Action Summit <a href="https://arxiv.org/pdf/2501.17805">report</a> - International AI Safety Report</li><li>YouGov <a href="https://today.yougov.com/technology/articles/51368-do-americans-think-ai-will-have-positive-or-negative-impact-society-artificial-intelligence-poll">article</a> - Do Americans think AI will have a positive or negative impact on society?</li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>I am joined by Tristan Williams and Felix de Simone to discuss their work on the potential of constituent communication, specifically in the context of AI legislation. These two worked as part of an AI Safety Camp team to understand whether or not it would be useful for more people to be sharing their experiences, concerns, and opinions with their government representative (hint, it is).</p><p>Check out the <a href="https://forum.effectivealtruism.org/posts/5oStggnYLGzomhvvn/talking-to-congress-can-constituents-contacting-their">blogpost</a> on their findings, "Talking to Congress: Can constituents contacting their legislator influence policy?" and the <a href="https://www.guidedtrack.com/programs/e7xnz7q/run">tool</a> they created!</p><p></p><ul><li>(01:53) - Introductions</li>
<li>(04:04) - Starting the project</li>
<li>(13:30) - Project overview</li>
<li>(16:36) - Understanding constituent communication</li>
<li>(28:50) - Literature review</li>
<li>(35:52) - Phase 2</li>
<li>(43:26) - Creating a tool for citizen engagement</li>
<li>(50:16) - Crafting your message</li>
<li>(59:40) - The game of advocacy</li>
<li>(01:15:19) - Difficulties on the project</li>
<li>(01:22:33) - Call to action</li>
<li>(01:32:30) - Outro</li>
</ul><br><strong>Links</strong><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.aisafety.camp/">AI Safety Camp</a></li><li><a href="https://pauseai.info/">Pause AI</a></li><li><a href="https://bluedot.org/">BlueDot Impact</a></li><li>TIME <a href="https://time.com/6972134/ai-lobbying-tech-policy-surge/">article</a> - There’s an AI Lobbying Frenzy in Washington. Big Tech Is Dominating</li><li>Congressional Management Foundation <a href="https://www.congressfoundation.org/storage/documents/CMF_Pubs/cwc-perceptions-of-citizen-advocacy.pdf">study</a> - Communicating with Congress: Perceptions of Citizen Advocacy on Capitol Hill</li><li>Congressional Management Foundation <a href="https://www.congressfoundation.org/storage/documents/CMF_Pubs/cmf_citizen_engagement_rebuilding_democratic_dialogue.pdf">study</a> - The Future of Citizen Engagement: Rebuilding the Democratic Dialogue</li><li>Tristan and Felix's <a href="https://forum.effectivealtruism.org/posts/5oStggnYLGzomhvvn/talking-to-congress-can-constituents-contacting-their">blogpost</a> - Talking to Congress: Can constituents contacting their legislator influence policy?</li><li>Wired <a href="https://www.wired.com/story/opengov-report-congress-constituent-communication/">article</a> - What It Takes to Make Congress Actually Listen</li><li>American Journal of Polical Science <a href="https://projects.iq.harvard.edu/files/cces/files/AnsolabehereKuriwaki_AJPS.pdf">article</a> - Congressional Representation: Accountability from the Constituent’s Perspective</li><li>Political Behavior <a href="https://www.jstor.org/stable/43653416">article</a> - Call Your Legislator: A Field Experimental Study of the Impact of a Constituency Mobilization Campaign on Legislative Voting</li><li>Guided Track <a href="https://www.guidedtrack.com/">website</a></li><li>The <a href="https://www.guidedtrack.com/programs/e7xnz7q/run">Tool</a></li><li>Holistic AI global regulatory <a href="https://tracker.holisticai.com/feed">tracker</a></li><li>White &amp; Case global regulatory <a href="https://www.whitecase.com/insight-our-thinking/ai-watch-global-regulatory-tracker">tracker</a></li><li>Steptoe US AI legislation <a href="https://www.steptoe.com/en/ai-legislative-tracker.html">tracker</a></li><li>Manatt US AIxHealth legislation <a href="https://www.manatt.com/insights/newsletters/health-highlights/manatt-health-health-ai-policy-tracker">tracker</a></li><li>Issue One <a href="https://issueone.org/articles/big-tech-spent-record-sums-on-lobbying-last-year/">article</a> - Big Tech Cozies Up to New Administration After Spending Record Sums on Lobbying Last Year</li><li>Verfassungsblog <a href="https://verfassungsblog.de/bigtechs-efforts-to-derail-the-ai-act/">article</a> - BigTech’s Efforts to Derail the AI Act</li><li>MIT Technology Review <a href="https://www.technologyreview.com/2025/01/21/1110260/openai-ups-its-lobbying-efforts-nearly-seven-fold/">article</a> - OpenAI has upped its lobbying efforts nearly sevenfold</li><li>Open Secrets <a href="https://www.opensecrets.org/federal-lobbying/issues/lobbyists?id=SCI&amp;year=2023">webpage</a> - Issue Profile: Science &amp; Technology</li><li>Statista <a href="https://www.statista.com/statistics/257344/top-lobbying-spenders-in-the-us/">data</a> - Leading lobbying spenders in the United States in 2024</li><li>Global Justice Now <a href="https://www.globaljustice.org.uk/resource/democracy-at-risk-in-davos-new-report-exposes-big-tech-lobbying-and-political-interference/">report</a> - Democracy at risk in Davos: new report exposes big tech lobbying and political interference</li><li>Ipsos <a href="https://www.ipsos.com/en-us/where-americans-stand-ai">article</a> - Where Americans stand on AI</li><li>AP-NORC <a href="https://apnorc.org/projects/there-is-bipartisan-concern-about-the-use-of-ai-in-the-2024-elections/">report</a> - There Is Bipartisan Concern About the Use of AI in the 2024 Elections</li><li>AI Action Summit <a href="https://arxiv.org/pdf/2501.17805">report</a> - International AI Safety Report</li><li>YouGov <a href="https://today.yougov.com/technology/articles/51368-do-americans-think-ai-will-have-positive-or-negative-impact-society-artificial-intelligence-poll">article</a> - Do Americans think AI will have a positive or negative impact on society?</li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 19 May 2025 09:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/a9294c9e/4bedf15c.mp3" length="89667924" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:image href="https://img.transistorcdn.com/WFoiFTnsFhIpc0tE9n_1vopS2cOD2ao3J5WpgnwiOpk/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8yMGJk/NzVlOWVhYzA0NzE4/MWFmMDM1MmY0OTZi/ZTFlMC5wbmc.jpg"/>
      <itunes:duration>5603</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>I am joined by Tristan Williams and Felix de Simone to discuss their work on the potential of constituent communication, specifically in the context of AI legislation. These two worked as part of an AI Safety Camp team to understand whether or not it would be useful for more people to be sharing their experiences, concerns, and opinions with their government representative (hint, it is).</p><p>Check out the <a href="https://forum.effectivealtruism.org/posts/5oStggnYLGzomhvvn/talking-to-congress-can-constituents-contacting-their">blogpost</a> on their findings, "Talking to Congress: Can constituents contacting their legislator influence policy?" and the <a href="https://www.guidedtrack.com/programs/e7xnz7q/run">tool</a> they created!</p><p></p><ul><li>(01:53) - Introductions</li>
<li>(04:04) - Starting the project</li>
<li>(13:30) - Project overview</li>
<li>(16:36) - Understanding constituent communication</li>
<li>(28:50) - Literature review</li>
<li>(35:52) - Phase 2</li>
<li>(43:26) - Creating a tool for citizen engagement</li>
<li>(50:16) - Crafting your message</li>
<li>(59:40) - The game of advocacy</li>
<li>(01:15:19) - Difficulties on the project</li>
<li>(01:22:33) - Call to action</li>
<li>(01:32:30) - Outro</li>
</ul><br><strong>Links</strong><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.aisafety.camp/">AI Safety Camp</a></li><li><a href="https://pauseai.info/">Pause AI</a></li><li><a href="https://bluedot.org/">BlueDot Impact</a></li><li>TIME <a href="https://time.com/6972134/ai-lobbying-tech-policy-surge/">article</a> - There’s an AI Lobbying Frenzy in Washington. Big Tech Is Dominating</li><li>Congressional Management Foundation <a href="https://www.congressfoundation.org/storage/documents/CMF_Pubs/cwc-perceptions-of-citizen-advocacy.pdf">study</a> - Communicating with Congress: Perceptions of Citizen Advocacy on Capitol Hill</li><li>Congressional Management Foundation <a href="https://www.congressfoundation.org/storage/documents/CMF_Pubs/cmf_citizen_engagement_rebuilding_democratic_dialogue.pdf">study</a> - The Future of Citizen Engagement: Rebuilding the Democratic Dialogue</li><li>Tristan and Felix's <a href="https://forum.effectivealtruism.org/posts/5oStggnYLGzomhvvn/talking-to-congress-can-constituents-contacting-their">blogpost</a> - Talking to Congress: Can constituents contacting their legislator influence policy?</li><li>Wired <a href="https://www.wired.com/story/opengov-report-congress-constituent-communication/">article</a> - What It Takes to Make Congress Actually Listen</li><li>American Journal of Polical Science <a href="https://projects.iq.harvard.edu/files/cces/files/AnsolabehereKuriwaki_AJPS.pdf">article</a> - Congressional Representation: Accountability from the Constituent’s Perspective</li><li>Political Behavior <a href="https://www.jstor.org/stable/43653416">article</a> - Call Your Legislator: A Field Experimental Study of the Impact of a Constituency Mobilization Campaign on Legislative Voting</li><li>Guided Track <a href="https://www.guidedtrack.com/">website</a></li><li>The <a href="https://www.guidedtrack.com/programs/e7xnz7q/run">Tool</a></li><li>Holistic AI global regulatory <a href="https://tracker.holisticai.com/feed">tracker</a></li><li>White &amp; Case global regulatory <a href="https://www.whitecase.com/insight-our-thinking/ai-watch-global-regulatory-tracker">tracker</a></li><li>Steptoe US AI legislation <a href="https://www.steptoe.com/en/ai-legislative-tracker.html">tracker</a></li><li>Manatt US AIxHealth legislation <a href="https://www.manatt.com/insights/newsletters/health-highlights/manatt-health-health-ai-policy-tracker">tracker</a></li><li>Issue One <a href="https://issueone.org/articles/big-tech-spent-record-sums-on-lobbying-last-year/">article</a> - Big Tech Cozies Up to New Administration After Spending Record Sums on Lobbying Last Year</li><li>Verfassungsblog <a href="https://verfassungsblog.de/bigtechs-efforts-to-derail-the-ai-act/">article</a> - BigTech’s Efforts to Derail the AI Act</li><li>MIT Technology Review <a href="https://www.technologyreview.com/2025/01/21/1110260/openai-ups-its-lobbying-efforts-nearly-seven-fold/">article</a> - OpenAI has upped its lobbying efforts nearly sevenfold</li><li>Open Secrets <a href="https://www.opensecrets.org/federal-lobbying/issues/lobbyists?id=SCI&amp;year=2023">webpage</a> - Issue Profile: Science &amp; Technology</li><li>Statista <a href="https://www.statista.com/statistics/257344/top-lobbying-spenders-in-the-us/">data</a> - Leading lobbying spenders in the United States in 2024</li><li>Global Justice Now <a href="https://www.globaljustice.org.uk/resource/democracy-at-risk-in-davos-new-report-exposes-big-tech-lobbying-and-political-interference/">report</a> - Democracy at risk in Davos: new report exposes big tech lobbying and political interference</li><li>Ipsos <a href="https://www.ipsos.com/en-us/where-americans-stand-ai">article</a> - Where Americans stand on AI</li><li>AP-NORC <a href="https://apnorc.org/projects/there-is-bipartisan-concern-about-the-use-of-ai-in-the-2024-elections/">report</a> - There Is Bipartisan Concern About the Use of AI in the 2024 Elections</li><li>AI Action Summit <a href="https://arxiv.org/pdf/2501.17805">report</a> - International AI Safety Report</li><li>YouGov <a href="https://today.yougov.com/technology/articles/51368-do-americans-think-ai-will-have-positive-or-negative-impact-society-artificial-intelligence-poll">article</a> - Do Americans think AI will have a positive or negative impact on society?</li></ul>]]>
      </itunes:summary>
      <itunes:keywords>technology, AI, democracy, constituency, advocacy, AI governance, United States AI, AI safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:chapters url="https://share.transistor.fm/s/a9294c9e/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>INTERVIEW: Scaling Democracy w/ (Dr.) Igor Krawczuk</title>
      <itunes:episode>19</itunes:episode>
      <podcast:episode>19</podcast:episode>
      <itunes:title>INTERVIEW: Scaling Democracy w/ (Dr.) Igor Krawczuk</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">81c52538-5af2-40aa-abd7-02c08088f43e</guid>
      <link>https://kairos.fm/intoaisafety/e019</link>
      <description>
        <![CDATA[<p>The <em>almost</em> Dr. Igor Krawczuk joins me for what is the equivalent of 4 of my previous episodes. We get into all the classics: eugenics, capitalism, philosophical toads... Need I say more?</p><p>If you're interested in connecting with Igor, head on over to his <a href="https://krawczuk.eu/">website</a>, or check out <a href="https://github.com/into-ai-safety/into-ai-safety.github.io/blob/master/_posts">placeholder for thesis</a> (it isn't published yet).</p><p>Because the full show notes have a whopping 115 additional links, I'll highlight some that I think are particularly worthwhile here:</p><ul><li>The best article you'll ever read on <a href="https://jacob-haimes.github.io/independent/Open-Source-AI-is-a-lie/">Open Source AI</a></li><li>The best article you'll ever read on <a href="https://www.odysseaninstitute.org/post/let-s-talk-about-emergence">emergence in ML</a></li><li>Kate Crawford's <a href="https://yalebooks.yale.edu/book/9780300264630/atlas-of-ai/"><em>Atlas of AI</em></a> (<a href="https://en.wikipedia.org/wiki/Atlas_of_AI">Wikipedia</a>)</li><li><a href="https://arxiv.org/abs/1911.01547">On the Measure of Intelligence</a></li><li>Thomas Piketty's <a href="https://www.hup.harvard.edu/books/9780674430006"><em>Capital in the Twenty-First Century</em></a> (<a href="https://en.wikipedia.org/wiki/Capital_in_the_Twenty-First_Century">Wikipedia</a>)</li><li>Yurii Nesterov's <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=2-ElBQAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;dq=info:JTiRBrZ_LZMJ:scholar.google.com&amp;ots=wnpRdsxjjv&amp;sig=1Oa-5P-zZZ_MX_2MFKv5cq2fx48#v=onepage&amp;q&amp;f=false"><em>Introductory Lectures on Convex Optimization</em></a></li></ul><p><strong>Chapters<br></strong></p><ul><li>(02:32) - Introducing Igor</li>
<li>(10:11) - Aside on EY, LW, EA, etc., a.k.a. lettersoup</li>
<li>(18:30) - Igor on AI alignment</li>
<li>(33:06) - "Open Source" in AI</li>
<li>(41:20) - The story of infinite riches and suffering</li>
<li>(59:11) - On AI threat models</li>
<li>(01:09:25) - Representation in AI</li>
<li>(01:15:00) - Hazard fishing</li>
<li>(01:18:52) - Intelligence and eugenics</li>
<li>(01:34:38) - Emergence</li>
<li>(01:48:19) - Considering externalities</li>
<li>(01:53:33) - The shape of an argument</li>
<li>(02:01:39) - More eugenics</li>
<li>(02:06:09) - I'm convinced, what now?</li>
<li>(02:18:03) - AIxBio (round ??)</li>
<li>(02:29:09) - On open release of models</li>
<li>(02:40:28) - Data and copyright</li>
<li>(02:44:09) - Scientific accessibility and bullshit</li>
<li>(02:53:04) - Igor's point of view</li>
<li>(02:57:20) - Outro</li>
</ul><p><strong><br>Links</strong></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. All references, including those only mentioned in the extended version of this episode, are included.</p><ul><li><a href="https://www.lighthousereports.com/suspicion-machines-methodology/">Suspicious Machines Methodology</a>, referred to as the "Rotterdam Lighthouse Report" in the episode</li><li><a href="https://www.epfl.ch/labs/lions/">LIONS Lab</a> at EPFL</li><li>The <a href="https://pbs.twimg.com/media/D53Q_MYW4AA-wRK.jpg">meme</a> that Igor references</li><li><a href="https://arxiv.org/abs/2401.01869">On the Hardness of Learning Under Symmetries</a></li><li><a href="https://uvagedl.github.io/">Course</a> on the concept of equivariant deep learning</li><li>Aside on EY/EA/etc.<ul><li>Sources on Eliezer Yudkowski<ul><li><a href="https://encyclopedia.pub/entry/33978">Scholarly Community Encyclopedia</a></li><li><a href="https://time.com/collection/time100-ai/6309037/eliezer-yudkowsky/">TIME100 AI</a></li><li>Yudkowski's personal <a href="https://www.yudkowsky.net/">website</a></li><li><a href="https://en.wikipedia.org/wiki/Eliezer_Yudkowsky">EY Wikipedia</a></li><li><a href="https://whatshouldiread.fandom.com/wiki/Eliezer_Yudkowsky#cite_note-1">A Very Literary Wiki</a> -TIME article: <a href="https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/">Pausing AI Developments Isn’t Enough. We Need to Shut it All Down</a> documenting EY's ruminations of bombing datacenters; this comes up later in the episode but is included here because it about EY.</li></ul></li><li><a href="https://www.lesswrong.com/">LessWrong</a><ul><li><a href="https://en.wikipedia.org/wiki/LessWrong">LW Wikipedia</a></li></ul></li><li><a href="https://intelligence.org/">MIRI</a></li><li>Coverage on Nick Bostrom (being a racist)<ul><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/28/nick-bostrom-controversial-future-of-humanity-institute-closure-longtermism-affective-altruism">‘Eugenics on steroids’: the toxic and contested legacy of Oxford’s Future of Humanity Institute</a></li><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/19/oxford-future-of-humanity-institute-closes">Oxford shuts down institute run by Elon Musk-backed philosopher</a></li></ul></li><li>Investigative <a href="https://markfuentes1.substack.com/p/emile-p-torress-history-of-dishonesty">piece</a> on Émile Torres</li><li><a href="https://dl.acm.org/doi/10.1145/3442188.3445922">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜</a></li><li>NY Times article: <a href="https://www.nytimes.com/2019/11/11/technology/artificial-intelligence-bias.html">We Teach A.I. Systems Everything, Including Our Biases</a></li><li>NY Times article: <a href="https://www.nytimes.com/2020/12/03/technology/google-researcher-timnit-gebru.html">Google Researcher Says She Was Fired Over Paper Highlighting Bias in A.I.</a></li><li>Timnit Gebru's <a href="https://en.wikipedia.org/wiki/Timnit_Gebru">Wikipedia</a></li><li><a href="https://firstmonday.org/ojs/index.php/fm/article/view/13636">The TESCREAL Bundle: Eugenics and the Promise of Utopia through Artificial General Intelligence</a></li><li>Sources on the environmental impact of LLMs<ul><li><a href="https://analyticsindiamag.com/the-environmental-impact-of-llms/">The Environmental Impact of LLMs</a></li><li><a href="https://tinyml.substack.com/p/the-cost-of-inference-running-the">The Cost of Inference: Running the Models</a></li><li><a href="https://arxiv.org/abs/1906.02243">Energy and Policy Considerations for Deep Learning in NLP</a></li><li><a href="https://weareyard.com/insights/the-carbon-impact-of-ai-vs-search-engines">The Carbon Impact of AI vs Search Engines</a></li></ul></li></ul></li><li><a href="https://www.science.org/doi/full/10.1126/science.abi7176?casa_token=2txe0r_jjhQAAAAA%3ALJa__HZL9COyj9EUpdILZdtnMKLyggfFe7Zpvv0tNze62rLO0CoQHCCJiXfruxUeBLj3YBZ33F8OOv0u">Filling Gaps in Trustworthy Development of AI </a>(Igor is an author on this one)</li><li><a href="https://www.hindawi.com/journals/complexity/2022/8210732/">A Computational Turn in Policy Process Studies: Coevolving Network Dynamics of Policy Change</a></li><li><a href="https://proceedings.neurips.cc/paper_files/paper/2020/file/7e05d6f828574fbc975a896b25bb011e-Paper.pdf">The Smoothed Possibility of Social Choice</a>, an intro in social choice theory and how it overlaps with ML</li><li>Relating to Dan Hendrycks<ul><li><a href="https://arxiv.org/abs/2303.16200">Natural Selection Favors AIs over Humans</a><ul><li>"One easy-to-digest source to highlight what he gets wrong [is] <a href="https://pressbooks.calstate.edu/explorationsbioanth2/chapter/17/">Social and Biopolitical Dimensions of Evolutionary Thinking</a>" -Igor</li></ul></li><li><a href="https://www.aisafetybook.com/">Introduction to AI Safety, Ethics, and Society</a>, recently published textbook</li><li>"<a href="https://arxiv.org/pdf/2306.12001#page=10.19">Source</a> to the section [of this paper] that makes Dan one of my favs from that crowd." -Igor</li><li><a href="https://twitter.com/DanHendrycks/status/1710312043503321141">Twitter post</a> referenced in the episode&lt;...</li></ul></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>The <em>almost</em> Dr. Igor Krawczuk joins me for what is the equivalent of 4 of my previous episodes. We get into all the classics: eugenics, capitalism, philosophical toads... Need I say more?</p><p>If you're interested in connecting with Igor, head on over to his <a href="https://krawczuk.eu/">website</a>, or check out <a href="https://github.com/into-ai-safety/into-ai-safety.github.io/blob/master/_posts">placeholder for thesis</a> (it isn't published yet).</p><p>Because the full show notes have a whopping 115 additional links, I'll highlight some that I think are particularly worthwhile here:</p><ul><li>The best article you'll ever read on <a href="https://jacob-haimes.github.io/independent/Open-Source-AI-is-a-lie/">Open Source AI</a></li><li>The best article you'll ever read on <a href="https://www.odysseaninstitute.org/post/let-s-talk-about-emergence">emergence in ML</a></li><li>Kate Crawford's <a href="https://yalebooks.yale.edu/book/9780300264630/atlas-of-ai/"><em>Atlas of AI</em></a> (<a href="https://en.wikipedia.org/wiki/Atlas_of_AI">Wikipedia</a>)</li><li><a href="https://arxiv.org/abs/1911.01547">On the Measure of Intelligence</a></li><li>Thomas Piketty's <a href="https://www.hup.harvard.edu/books/9780674430006"><em>Capital in the Twenty-First Century</em></a> (<a href="https://en.wikipedia.org/wiki/Capital_in_the_Twenty-First_Century">Wikipedia</a>)</li><li>Yurii Nesterov's <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=2-ElBQAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;dq=info:JTiRBrZ_LZMJ:scholar.google.com&amp;ots=wnpRdsxjjv&amp;sig=1Oa-5P-zZZ_MX_2MFKv5cq2fx48#v=onepage&amp;q&amp;f=false"><em>Introductory Lectures on Convex Optimization</em></a></li></ul><p><strong>Chapters<br></strong></p><ul><li>(02:32) - Introducing Igor</li>
<li>(10:11) - Aside on EY, LW, EA, etc., a.k.a. lettersoup</li>
<li>(18:30) - Igor on AI alignment</li>
<li>(33:06) - "Open Source" in AI</li>
<li>(41:20) - The story of infinite riches and suffering</li>
<li>(59:11) - On AI threat models</li>
<li>(01:09:25) - Representation in AI</li>
<li>(01:15:00) - Hazard fishing</li>
<li>(01:18:52) - Intelligence and eugenics</li>
<li>(01:34:38) - Emergence</li>
<li>(01:48:19) - Considering externalities</li>
<li>(01:53:33) - The shape of an argument</li>
<li>(02:01:39) - More eugenics</li>
<li>(02:06:09) - I'm convinced, what now?</li>
<li>(02:18:03) - AIxBio (round ??)</li>
<li>(02:29:09) - On open release of models</li>
<li>(02:40:28) - Data and copyright</li>
<li>(02:44:09) - Scientific accessibility and bullshit</li>
<li>(02:53:04) - Igor's point of view</li>
<li>(02:57:20) - Outro</li>
</ul><p><strong><br>Links</strong></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. All references, including those only mentioned in the extended version of this episode, are included.</p><ul><li><a href="https://www.lighthousereports.com/suspicion-machines-methodology/">Suspicious Machines Methodology</a>, referred to as the "Rotterdam Lighthouse Report" in the episode</li><li><a href="https://www.epfl.ch/labs/lions/">LIONS Lab</a> at EPFL</li><li>The <a href="https://pbs.twimg.com/media/D53Q_MYW4AA-wRK.jpg">meme</a> that Igor references</li><li><a href="https://arxiv.org/abs/2401.01869">On the Hardness of Learning Under Symmetries</a></li><li><a href="https://uvagedl.github.io/">Course</a> on the concept of equivariant deep learning</li><li>Aside on EY/EA/etc.<ul><li>Sources on Eliezer Yudkowski<ul><li><a href="https://encyclopedia.pub/entry/33978">Scholarly Community Encyclopedia</a></li><li><a href="https://time.com/collection/time100-ai/6309037/eliezer-yudkowsky/">TIME100 AI</a></li><li>Yudkowski's personal <a href="https://www.yudkowsky.net/">website</a></li><li><a href="https://en.wikipedia.org/wiki/Eliezer_Yudkowsky">EY Wikipedia</a></li><li><a href="https://whatshouldiread.fandom.com/wiki/Eliezer_Yudkowsky#cite_note-1">A Very Literary Wiki</a> -TIME article: <a href="https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/">Pausing AI Developments Isn’t Enough. We Need to Shut it All Down</a> documenting EY's ruminations of bombing datacenters; this comes up later in the episode but is included here because it about EY.</li></ul></li><li><a href="https://www.lesswrong.com/">LessWrong</a><ul><li><a href="https://en.wikipedia.org/wiki/LessWrong">LW Wikipedia</a></li></ul></li><li><a href="https://intelligence.org/">MIRI</a></li><li>Coverage on Nick Bostrom (being a racist)<ul><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/28/nick-bostrom-controversial-future-of-humanity-institute-closure-longtermism-affective-altruism">‘Eugenics on steroids’: the toxic and contested legacy of Oxford’s Future of Humanity Institute</a></li><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/19/oxford-future-of-humanity-institute-closes">Oxford shuts down institute run by Elon Musk-backed philosopher</a></li></ul></li><li>Investigative <a href="https://markfuentes1.substack.com/p/emile-p-torress-history-of-dishonesty">piece</a> on Émile Torres</li><li><a href="https://dl.acm.org/doi/10.1145/3442188.3445922">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜</a></li><li>NY Times article: <a href="https://www.nytimes.com/2019/11/11/technology/artificial-intelligence-bias.html">We Teach A.I. Systems Everything, Including Our Biases</a></li><li>NY Times article: <a href="https://www.nytimes.com/2020/12/03/technology/google-researcher-timnit-gebru.html">Google Researcher Says She Was Fired Over Paper Highlighting Bias in A.I.</a></li><li>Timnit Gebru's <a href="https://en.wikipedia.org/wiki/Timnit_Gebru">Wikipedia</a></li><li><a href="https://firstmonday.org/ojs/index.php/fm/article/view/13636">The TESCREAL Bundle: Eugenics and the Promise of Utopia through Artificial General Intelligence</a></li><li>Sources on the environmental impact of LLMs<ul><li><a href="https://analyticsindiamag.com/the-environmental-impact-of-llms/">The Environmental Impact of LLMs</a></li><li><a href="https://tinyml.substack.com/p/the-cost-of-inference-running-the">The Cost of Inference: Running the Models</a></li><li><a href="https://arxiv.org/abs/1906.02243">Energy and Policy Considerations for Deep Learning in NLP</a></li><li><a href="https://weareyard.com/insights/the-carbon-impact-of-ai-vs-search-engines">The Carbon Impact of AI vs Search Engines</a></li></ul></li></ul></li><li><a href="https://www.science.org/doi/full/10.1126/science.abi7176?casa_token=2txe0r_jjhQAAAAA%3ALJa__HZL9COyj9EUpdILZdtnMKLyggfFe7Zpvv0tNze62rLO0CoQHCCJiXfruxUeBLj3YBZ33F8OOv0u">Filling Gaps in Trustworthy Development of AI </a>(Igor is an author on this one)</li><li><a href="https://www.hindawi.com/journals/complexity/2022/8210732/">A Computational Turn in Policy Process Studies: Coevolving Network Dynamics of Policy Change</a></li><li><a href="https://proceedings.neurips.cc/paper_files/paper/2020/file/7e05d6f828574fbc975a896b25bb011e-Paper.pdf">The Smoothed Possibility of Social Choice</a>, an intro in social choice theory and how it overlaps with ML</li><li>Relating to Dan Hendrycks<ul><li><a href="https://arxiv.org/abs/2303.16200">Natural Selection Favors AIs over Humans</a><ul><li>"One easy-to-digest source to highlight what he gets wrong [is] <a href="https://pressbooks.calstate.edu/explorationsbioanth2/chapter/17/">Social and Biopolitical Dimensions of Evolutionary Thinking</a>" -Igor</li></ul></li><li><a href="https://www.aisafetybook.com/">Introduction to AI Safety, Ethics, and Society</a>, recently published textbook</li><li>"<a href="https://arxiv.org/pdf/2306.12001#page=10.19">Source</a> to the section [of this paper] that makes Dan one of my favs from that crowd." -Igor</li><li><a href="https://twitter.com/DanHendrycks/status/1710312043503321141">Twitter post</a> referenced in the episode&lt;...</li></ul></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 03 Jun 2024 10:11:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/b8225038/75bd8800.mp3" length="171643841" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>10726</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>The <em>almost</em> Dr. Igor Krawczuk joins me for what is the equivalent of 4 of my previous episodes. We get into all the classics: eugenics, capitalism, philosophical toads... Need I say more?</p><p>If you're interested in connecting with Igor, head on over to his <a href="https://krawczuk.eu/">website</a>, or check out <a href="https://github.com/into-ai-safety/into-ai-safety.github.io/blob/master/_posts">placeholder for thesis</a> (it isn't published yet).</p><p>Because the full show notes have a whopping 115 additional links, I'll highlight some that I think are particularly worthwhile here:</p><ul><li>The best article you'll ever read on <a href="https://jacob-haimes.github.io/independent/Open-Source-AI-is-a-lie/">Open Source AI</a></li><li>The best article you'll ever read on <a href="https://www.odysseaninstitute.org/post/let-s-talk-about-emergence">emergence in ML</a></li><li>Kate Crawford's <a href="https://yalebooks.yale.edu/book/9780300264630/atlas-of-ai/"><em>Atlas of AI</em></a> (<a href="https://en.wikipedia.org/wiki/Atlas_of_AI">Wikipedia</a>)</li><li><a href="https://arxiv.org/abs/1911.01547">On the Measure of Intelligence</a></li><li>Thomas Piketty's <a href="https://www.hup.harvard.edu/books/9780674430006"><em>Capital in the Twenty-First Century</em></a> (<a href="https://en.wikipedia.org/wiki/Capital_in_the_Twenty-First_Century">Wikipedia</a>)</li><li>Yurii Nesterov's <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=2-ElBQAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;dq=info:JTiRBrZ_LZMJ:scholar.google.com&amp;ots=wnpRdsxjjv&amp;sig=1Oa-5P-zZZ_MX_2MFKv5cq2fx48#v=onepage&amp;q&amp;f=false"><em>Introductory Lectures on Convex Optimization</em></a></li></ul><p><strong>Chapters<br></strong></p><ul><li>(02:32) - Introducing Igor</li>
<li>(10:11) - Aside on EY, LW, EA, etc., a.k.a. lettersoup</li>
<li>(18:30) - Igor on AI alignment</li>
<li>(33:06) - "Open Source" in AI</li>
<li>(41:20) - The story of infinite riches and suffering</li>
<li>(59:11) - On AI threat models</li>
<li>(01:09:25) - Representation in AI</li>
<li>(01:15:00) - Hazard fishing</li>
<li>(01:18:52) - Intelligence and eugenics</li>
<li>(01:34:38) - Emergence</li>
<li>(01:48:19) - Considering externalities</li>
<li>(01:53:33) - The shape of an argument</li>
<li>(02:01:39) - More eugenics</li>
<li>(02:06:09) - I'm convinced, what now?</li>
<li>(02:18:03) - AIxBio (round ??)</li>
<li>(02:29:09) - On open release of models</li>
<li>(02:40:28) - Data and copyright</li>
<li>(02:44:09) - Scientific accessibility and bullshit</li>
<li>(02:53:04) - Igor's point of view</li>
<li>(02:57:20) - Outro</li>
</ul><p><strong><br>Links</strong></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. All references, including those only mentioned in the extended version of this episode, are included.</p><ul><li><a href="https://www.lighthousereports.com/suspicion-machines-methodology/">Suspicious Machines Methodology</a>, referred to as the "Rotterdam Lighthouse Report" in the episode</li><li><a href="https://www.epfl.ch/labs/lions/">LIONS Lab</a> at EPFL</li><li>The <a href="https://pbs.twimg.com/media/D53Q_MYW4AA-wRK.jpg">meme</a> that Igor references</li><li><a href="https://arxiv.org/abs/2401.01869">On the Hardness of Learning Under Symmetries</a></li><li><a href="https://uvagedl.github.io/">Course</a> on the concept of equivariant deep learning</li><li>Aside on EY/EA/etc.<ul><li>Sources on Eliezer Yudkowski<ul><li><a href="https://encyclopedia.pub/entry/33978">Scholarly Community Encyclopedia</a></li><li><a href="https://time.com/collection/time100-ai/6309037/eliezer-yudkowsky/">TIME100 AI</a></li><li>Yudkowski's personal <a href="https://www.yudkowsky.net/">website</a></li><li><a href="https://en.wikipedia.org/wiki/Eliezer_Yudkowsky">EY Wikipedia</a></li><li><a href="https://whatshouldiread.fandom.com/wiki/Eliezer_Yudkowsky#cite_note-1">A Very Literary Wiki</a> -TIME article: <a href="https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/">Pausing AI Developments Isn’t Enough. We Need to Shut it All Down</a> documenting EY's ruminations of bombing datacenters; this comes up later in the episode but is included here because it about EY.</li></ul></li><li><a href="https://www.lesswrong.com/">LessWrong</a><ul><li><a href="https://en.wikipedia.org/wiki/LessWrong">LW Wikipedia</a></li></ul></li><li><a href="https://intelligence.org/">MIRI</a></li><li>Coverage on Nick Bostrom (being a racist)<ul><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/28/nick-bostrom-controversial-future-of-humanity-institute-closure-longtermism-affective-altruism">‘Eugenics on steroids’: the toxic and contested legacy of Oxford’s Future of Humanity Institute</a></li><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/19/oxford-future-of-humanity-institute-closes">Oxford shuts down institute run by Elon Musk-backed philosopher</a></li></ul></li><li>Investigative <a href="https://markfuentes1.substack.com/p/emile-p-torress-history-of-dishonesty">piece</a> on Émile Torres</li><li><a href="https://dl.acm.org/doi/10.1145/3442188.3445922">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜</a></li><li>NY Times article: <a href="https://www.nytimes.com/2019/11/11/technology/artificial-intelligence-bias.html">We Teach A.I. Systems Everything, Including Our Biases</a></li><li>NY Times article: <a href="https://www.nytimes.com/2020/12/03/technology/google-researcher-timnit-gebru.html">Google Researcher Says She Was Fired Over Paper Highlighting Bias in A.I.</a></li><li>Timnit Gebru's <a href="https://en.wikipedia.org/wiki/Timnit_Gebru">Wikipedia</a></li><li><a href="https://firstmonday.org/ojs/index.php/fm/article/view/13636">The TESCREAL Bundle: Eugenics and the Promise of Utopia through Artificial General Intelligence</a></li><li>Sources on the environmental impact of LLMs<ul><li><a href="https://analyticsindiamag.com/the-environmental-impact-of-llms/">The Environmental Impact of LLMs</a></li><li><a href="https://tinyml.substack.com/p/the-cost-of-inference-running-the">The Cost of Inference: Running the Models</a></li><li><a href="https://arxiv.org/abs/1906.02243">Energy and Policy Considerations for Deep Learning in NLP</a></li><li><a href="https://weareyard.com/insights/the-carbon-impact-of-ai-vs-search-engines">The Carbon Impact of AI vs Search Engines</a></li></ul></li></ul></li><li><a href="https://www.science.org/doi/full/10.1126/science.abi7176?casa_token=2txe0r_jjhQAAAAA%3ALJa__HZL9COyj9EUpdILZdtnMKLyggfFe7Zpvv0tNze62rLO0CoQHCCJiXfruxUeBLj3YBZ33F8OOv0u">Filling Gaps in Trustworthy Development of AI </a>(Igor is an author on this one)</li><li><a href="https://www.hindawi.com/journals/complexity/2022/8210732/">A Computational Turn in Policy Process Studies: Coevolving Network Dynamics of Policy Change</a></li><li><a href="https://proceedings.neurips.cc/paper_files/paper/2020/file/7e05d6f828574fbc975a896b25bb011e-Paper.pdf">The Smoothed Possibility of Social Choice</a>, an intro in social choice theory and how it overlaps with ML</li><li>Relating to Dan Hendrycks<ul><li><a href="https://arxiv.org/abs/2303.16200">Natural Selection Favors AIs over Humans</a><ul><li>"One easy-to-digest source to highlight what he gets wrong [is] <a href="https://pressbooks.calstate.edu/explorationsbioanth2/chapter/17/">Social and Biopolitical Dimensions of Evolutionary Thinking</a>" -Igor</li></ul></li><li><a href="https://www.aisafetybook.com/">Introduction to AI Safety, Ethics, and Society</a>, recently published textbook</li><li>"<a href="https://arxiv.org/pdf/2306.12001#page=10.19">Source</a> to the section [of this paper] that makes Dan one of my favs from that crowd." -Igor</li><li><a href="https://twitter.com/DanHendrycks/status/1710312043503321141">Twitter post</a> referenced in the episode&lt;...</li></ul></li></ul>]]>
      </itunes:summary>
      <itunes:keywords>AI, machine learning, democracy, AI safety</itunes:keywords>
      <itunes:explicit>Yes</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://krawczuk.eu" img="https://img.transistorcdn.com/pNx-3anl_InsAyWU5Hdrt7WvpZbrq2_twy1JYiyST3I/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9jZTE4/YTI3ZTgyYjY4YzFh/OTY0NDkzNmZjZTUz/YjVkYi5qcGVn.jpg">Dr. Igor Krawczuk</podcast:person>
      <podcast:chapters url="https://share.transistor.fm/s/b8225038/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (3)</title>
      <itunes:episode>18</itunes:episode>
      <podcast:episode>18</podcast:episode>
      <itunes:title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (3)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">0ca91074-f16c-428a-9408-550bfb7ceb4f</guid>
      <link>https://kairos.fm/intoaisafety/e018</link>
      <description>
        <![CDATA[<p>As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.</p><p>As you may have ascertained from the previous two segments of the interview, Dr. Park cofounded <a href="https://www.stakeout.ai">StakeOut.AI</a> along with Harry Luk and one other cofounder whose name has been removed due to requirements of her current position. The non-profit had a simple but important mission: make the adoption of AI technology go well, for humanity, but unfortunately, StakeOut.AI had to dissolve in late February of 2024 because no granter would fund them. Although it certainly is disappointing that the organization is no longer functioning, all three cofounders continue to contribute positively towards improving our world in their current roles.</p><p>If you would like to investigate further into Dr. Park's work, view his <a href="https://scholar.harvard.edu/pspark">website</a>, <a href="https://scholar.google.com/citations?user=5lMAPEoAAAAJ&amp;hl=en">Google Scholar</a>, or follow him on <a href="https://twitter.com/dr_park_phd">Twitter</a></p><p>00:00:54 ❙ Intro<br>00:02:41 ❙ Rapid development<br>00:08:25 ❙ Provable safety, safety factors, &amp; CSAM<br>00:18:50 ❙ Litigation<br>00:23:06 ❙ Open/Closed Source<br>00:38:52 ❙ AIxBio<br>00:47:50 ❙ Scientific rigor in AI<br>00:56:22 ❙ AI deception<br>01:02:45 ❙ No takesies-backsies<br>01:08:22 ❙ StakeOut.AI's start<br>01:12:53 ❙ Sustainability &amp; Agency<br>01:18:21 ❙ "I'm sold, next steps?" -you<br>01:23:53 ❙ Lessons from the amazing Spiderman<br>01:33:15 ❙ "I'm ready to switch careers, next steps?" -you<br>01:40:00 ❙ The most important question<br>01:41:11 ❙ Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://pauseai.info">Pause AI</a></li><li><a href="https://futureoflife.org/wp-content/uploads/2023/11/FLI_Governance_Scorecard_and_Framework.pdf">AI Governance Scorecard</a> (go to Pg. 3)</li><li><a href="https://civitai.com">CIVITAI</a><ul><li><a href="https://www.404media.co/a16z-funded-ai-platform-generated-images-that-could-be-categorized-as-child-pornography-leaked-documents-show/">Article on CIVITAI and CSAM</a></li></ul></li><li><a href="https://www.judiciary.senate.gov/protecting-children-online">Senate Hearing: Protecting Children Online</a><ul><li><a href="https://www.pbs.org/newshour/politics/watch-live-ceos-of-meta-tiktok-x-and-other-social-media-companies-testify-in-senate-hearing">PBS Newshour Coverage</a></li></ul></li><li><a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work</a></li><li>Open Source/Weights/Release/Interpretation<ul><li><a href="https://opensource.org">Open Source Initiative</a><ul><li><a href="https://opensource.org/history">History of the OSI</a></li><li><a href="https://opensource.org/blog/metas-llama-2-license-is-not-open-source">Meta’s LLaMa 2 license is not Open Source</a></li></ul></li><li><a href="https://opensourceconnections.com/blog/2023/07/19/is-llama-2-open-source-no-and-perhaps-we-need-a-new-definition-of-open/">Is Llama 2 open source? No – and perhaps we need a new definition of open…</a></li><li><a href="https://www.apache.org/licenses/LICENSE-2.0">Apache License, Version 2.0</a></li><li><a href="https://www.3blue1brown.com/topics/neural-networks">3Blue1Brown: Neural Networks</a></li><li><a href="https://dl.acm.org/doi/10.1145/3571884.3604316">Opening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generators</a><ul><li>The online <a href="https://opening-up-chatgpt.github.io">table</a></li></ul></li></ul></li><li><a href="https://www.signal.org">Signal</a></li><li><a href="https://huggingface.co/bigscience/bloomz">Bloomz</a> model on HuggingFace</li><li><a href="https://mistral.ai">Mistral</a> website</li><li>NASA Tragedies<ul><li><a href="https://en.wikipedia.org/wiki/Space_Shuttle_Challenger_disaster">Challenger disaster</a> on Wikipedia</li><li><a href="https://en.wikipedia.org/wiki/Space_Shuttle_Columbia_disaster">Columbia disaster</a> on Wikipedia</li></ul></li><li>AIxBio Risk<ul><li><a href="https://www.nature.com/articles/s42256-022-00465-9">Dual use of artificial-intelligence-powered drug discovery</a></li><li><a href="https://arxiv.org/abs/2306.03809">Can large language models democratize access to dual-use biotechnology?</a></li><li><a href="https://www.governance.ai/research-paper/open-sourcing-highly-capable-foundation-models">Open-Sourcing Highly Capable Foundation Models</a> <em>(sadly, I can't rename the article...)</em></li><li><a href="https://1a3orn.com/sub/essays-propaganda-or-science.html">Propaganda or Science: Open Source AI and Bioterrorism Risk</a></li><li><a href="https://ineffectivealtruismblog.com/2024/03/09/exaggerating-the-risks-part-14-biorisk-from-llms/">Exaggerating the risks (Part 15: Biorisk from LLMs)</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will releasing the weights of future large language models grant widespread access to pandemic agents?</a></li><li><a href="https://crfm.stanford.edu/open-fms/">On the Societal Impact of Open Foundation Models</a><ul><li><a href="https://hai.stanford.edu/sites/default/files/2023-12/Governing-Open-Foundation-Models.pdf">Policy brief</a></li></ul></li></ul></li><li><a href="https://www.apartresearch.com">Apart Research</a></li><li><a href="https://www.science.org">Science</a></li><li>Cicero<ul><li><a href="https://www.science.org/doi/10.1126/science.ade9097">Human-level play in the game of Diplomacy by combining language models with strategic reasoning</a></li><li><a href="https://ai.meta.com/research/cicero/">Cicero</a> webpage</li><li><a href="https://arxiv.org/abs/2308.14752">AI Deception: A Survey of Examples, Risks, and Potential Solutions</a></li></ul></li><li><a href="https://demos.co.uk/research/open-sourcing-the-ai-revolution-framing-the-debate-on-open-source-artificial-intelligence-and-regulation/">Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation</a></li><li><a href="https://aisafety.camp">AI Safety Camp</a></li><li><a href="https://www.patreon.com/IntoAISafety">Into AI Safety Patreon</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.</p><p>As you may have ascertained from the previous two segments of the interview, Dr. Park cofounded <a href="https://www.stakeout.ai">StakeOut.AI</a> along with Harry Luk and one other cofounder whose name has been removed due to requirements of her current position. The non-profit had a simple but important mission: make the adoption of AI technology go well, for humanity, but unfortunately, StakeOut.AI had to dissolve in late February of 2024 because no granter would fund them. Although it certainly is disappointing that the organization is no longer functioning, all three cofounders continue to contribute positively towards improving our world in their current roles.</p><p>If you would like to investigate further into Dr. Park's work, view his <a href="https://scholar.harvard.edu/pspark">website</a>, <a href="https://scholar.google.com/citations?user=5lMAPEoAAAAJ&amp;hl=en">Google Scholar</a>, or follow him on <a href="https://twitter.com/dr_park_phd">Twitter</a></p><p>00:00:54 ❙ Intro<br>00:02:41 ❙ Rapid development<br>00:08:25 ❙ Provable safety, safety factors, &amp; CSAM<br>00:18:50 ❙ Litigation<br>00:23:06 ❙ Open/Closed Source<br>00:38:52 ❙ AIxBio<br>00:47:50 ❙ Scientific rigor in AI<br>00:56:22 ❙ AI deception<br>01:02:45 ❙ No takesies-backsies<br>01:08:22 ❙ StakeOut.AI's start<br>01:12:53 ❙ Sustainability &amp; Agency<br>01:18:21 ❙ "I'm sold, next steps?" -you<br>01:23:53 ❙ Lessons from the amazing Spiderman<br>01:33:15 ❙ "I'm ready to switch careers, next steps?" -you<br>01:40:00 ❙ The most important question<br>01:41:11 ❙ Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://pauseai.info">Pause AI</a></li><li><a href="https://futureoflife.org/wp-content/uploads/2023/11/FLI_Governance_Scorecard_and_Framework.pdf">AI Governance Scorecard</a> (go to Pg. 3)</li><li><a href="https://civitai.com">CIVITAI</a><ul><li><a href="https://www.404media.co/a16z-funded-ai-platform-generated-images-that-could-be-categorized-as-child-pornography-leaked-documents-show/">Article on CIVITAI and CSAM</a></li></ul></li><li><a href="https://www.judiciary.senate.gov/protecting-children-online">Senate Hearing: Protecting Children Online</a><ul><li><a href="https://www.pbs.org/newshour/politics/watch-live-ceos-of-meta-tiktok-x-and-other-social-media-companies-testify-in-senate-hearing">PBS Newshour Coverage</a></li></ul></li><li><a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work</a></li><li>Open Source/Weights/Release/Interpretation<ul><li><a href="https://opensource.org">Open Source Initiative</a><ul><li><a href="https://opensource.org/history">History of the OSI</a></li><li><a href="https://opensource.org/blog/metas-llama-2-license-is-not-open-source">Meta’s LLaMa 2 license is not Open Source</a></li></ul></li><li><a href="https://opensourceconnections.com/blog/2023/07/19/is-llama-2-open-source-no-and-perhaps-we-need-a-new-definition-of-open/">Is Llama 2 open source? No – and perhaps we need a new definition of open…</a></li><li><a href="https://www.apache.org/licenses/LICENSE-2.0">Apache License, Version 2.0</a></li><li><a href="https://www.3blue1brown.com/topics/neural-networks">3Blue1Brown: Neural Networks</a></li><li><a href="https://dl.acm.org/doi/10.1145/3571884.3604316">Opening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generators</a><ul><li>The online <a href="https://opening-up-chatgpt.github.io">table</a></li></ul></li></ul></li><li><a href="https://www.signal.org">Signal</a></li><li><a href="https://huggingface.co/bigscience/bloomz">Bloomz</a> model on HuggingFace</li><li><a href="https://mistral.ai">Mistral</a> website</li><li>NASA Tragedies<ul><li><a href="https://en.wikipedia.org/wiki/Space_Shuttle_Challenger_disaster">Challenger disaster</a> on Wikipedia</li><li><a href="https://en.wikipedia.org/wiki/Space_Shuttle_Columbia_disaster">Columbia disaster</a> on Wikipedia</li></ul></li><li>AIxBio Risk<ul><li><a href="https://www.nature.com/articles/s42256-022-00465-9">Dual use of artificial-intelligence-powered drug discovery</a></li><li><a href="https://arxiv.org/abs/2306.03809">Can large language models democratize access to dual-use biotechnology?</a></li><li><a href="https://www.governance.ai/research-paper/open-sourcing-highly-capable-foundation-models">Open-Sourcing Highly Capable Foundation Models</a> <em>(sadly, I can't rename the article...)</em></li><li><a href="https://1a3orn.com/sub/essays-propaganda-or-science.html">Propaganda or Science: Open Source AI and Bioterrorism Risk</a></li><li><a href="https://ineffectivealtruismblog.com/2024/03/09/exaggerating-the-risks-part-14-biorisk-from-llms/">Exaggerating the risks (Part 15: Biorisk from LLMs)</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will releasing the weights of future large language models grant widespread access to pandemic agents?</a></li><li><a href="https://crfm.stanford.edu/open-fms/">On the Societal Impact of Open Foundation Models</a><ul><li><a href="https://hai.stanford.edu/sites/default/files/2023-12/Governing-Open-Foundation-Models.pdf">Policy brief</a></li></ul></li></ul></li><li><a href="https://www.apartresearch.com">Apart Research</a></li><li><a href="https://www.science.org">Science</a></li><li>Cicero<ul><li><a href="https://www.science.org/doi/10.1126/science.ade9097">Human-level play in the game of Diplomacy by combining language models with strategic reasoning</a></li><li><a href="https://ai.meta.com/research/cicero/">Cicero</a> webpage</li><li><a href="https://arxiv.org/abs/2308.14752">AI Deception: A Survey of Examples, Risks, and Potential Solutions</a></li></ul></li><li><a href="https://demos.co.uk/research/open-sourcing-the-ai-revolution-framing-the-debate-on-open-source-artificial-intelligence-and-regulation/">Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation</a></li><li><a href="https://aisafety.camp">AI Safety Camp</a></li><li><a href="https://www.patreon.com/IntoAISafety">Into AI Safety Patreon</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 25 Mar 2024 09:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/8c7eea4d/fac829ef.mp3" length="97914928" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>6120</itunes:duration>
      <itunes:summary>As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.As you may have ascertained from the previous two segments of the interview, Dr. Park cofounded StakeOut.AI along with Harry Luk and one other cofounder whose name has been removed due to requirements of her current position. The non-profit had a simple but important mission: make the adoption of AI technology go well, for humanity, but unfortunately, StakeOut.AI had to dissolve in late February of 2024 because no granter would fund them. Although it certainly is disappointing that the organization is no longer functioning, all three cofounders continue to contribute positively towards improving our world in their current roles.If you would like to investigate further into Dr. Park's work, view his website, Google Scholar, or follow him on Twitter00:00:54 ❙ Intro00:02:41 ❙ Rapid development00:08:25 ❙ Provable safety, safety factors, &amp;amp; CSAM00:18:50 ❙ Litigation00:23:06 ❙ Open/Closed Source00:38:52 ❙ AIxBio00:47:50 ❙ Scientific rigor in AI00:56:22 ❙ AI deception01:02:45 ❙ No takesies-backsies01:08:22 ❙ StakeOut.AI's start01:12:53 ❙ Sustainability &amp;amp; Agency01:18:21 ❙ "I'm sold, next steps?" -you01:23:53 ❙ Lessons from the amazing Spiderman01:33:15 ❙ "I'm ready to switch careers, next steps?" -you01:40:00 ❙ The most important question01:41:11 ❙ OutroLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.StakeOut.AIPause AIAI Governance Scorecard (go to Pg. 3)CIVITAIArticle on CIVITAI and CSAMSenate Hearing: Protecting Children OnlinePBS Newshour CoverageThe Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted WorkOpen Source/Weights/Release/InterpretationOpen Source InitiativeHistory of the OSIMeta’s LLaMa 2 license is not Open SourceIs Llama 2 open source? No – and perhaps we need a new definition of open…Apache License, Version 2.03Blue1Brown: Neural NetworksOpening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generatorsThe online tableSignalBloomz model on HuggingFaceMistral websiteNASA TragediesChallenger disaster on WikipediaColumbia disaster on WikipediaAIxBio RiskDual use of artificial-intelligence-powered drug discoveryCan large language models democratize access to dual-use biotechnology?Open-Sourcing Highly Capable Foundation Models (sadly, I can't rename the article...)Propaganda or Science: Open Source AI and Bioterrorism RiskExaggerating the risks (Part 15: Biorisk from LLMs)Will releasing the weights of future large language models grant widespread access to pandemic agents?On the Societal Impact of Open Foundation ModelsPolicy briefApart ResearchScienceCiceroHuman-level play in the game of Diplomacy by combining language models with strategic reasoningCicero webpageAI Deception: A Survey of Examples, Risks, and Potential SolutionsOpen Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulationAI Safety CampInto AI Safety Patreon</itunes:summary>
      <itunes:subtitle>As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.As you may have ascertained from the prev</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://scholar.harvard.edu/pspark" img="https://img.transistorcdn.com/uwPq3iaX1my6lU3bJjoP7cJ1SyqboZxZpJmB5Sw8bDc/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9mNjIz/ZmM0ZGJlMDU0NDZj/NDQ4NGVjNjE0YjFm/NDE1MC5qcGc.jpg">Dr. Peter S. Park</podcast:person>
    </item>
    <item>
      <title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (2)</title>
      <itunes:episode>17</itunes:episode>
      <podcast:episode>17</podcast:episode>
      <itunes:title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (2)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">ecb55f43-ac84-4293-b93c-0b9562b236a4</guid>
      <link>https://kairos.fm/intoaisafety/e017</link>
      <description>
        <![CDATA[<p>Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of <a href="https://www.stakeout.ai">StakeOut.AI</a>, a non-profit focused on making AI go well <em>for humans</em>, along with Harry Luk and one other individual, whose name has been removed due to requirements of her current position.</p><p>In addition to the normal links, I wanted to include the links to the petitions that Dr. Park mentions during the podcast. Note that the nonprofit which began these petitions, StakeOut.AI, has been dissolved.<br><a href="https://www.change.org/p/right-ai-laws-to-right-our-future-support-artificial-intelligence-safety-regulations-now">Right AI Laws, to Right Our Future: Support Artificial Intelligence Safety Regulations Now</a><br><a href="https://www.change.org/p/is-deepfake-illegal-not-yet-ban-deepfakes-to-protect-your-family-demand-deepfake-laws">Is Deepfake Illegal? Not Yet! Ban Deepfakes to Protect Your Family &amp; Demand Deepfake Laws</a><br><a href="https://www.change.org/p/ban-superintelligence-stop-ai-driven-human-extinction-risk">Ban Superintelligence: Stop AI-Driven Human Extinction Risk </a></p><p>00:00:54 - Intro<br>00:02:34 - Battleground 1: Copyright<br>00:06:28 - Battleground 2: Moral Critique of AI Collaborationists<br>00:08:15 - Rich Sutton<br>00:20:41 - OpenAI <em>Drama</em><br>00:34:28 - Battleground 3: Contract Negotiations for AI Ban Clauses<br>00:37:57 - Tesla, Autopilot, and FSD<br>00:40:02 - Recycling<br>00:47:40 - Battleground 4: New Laws and Policies<br>00:50:00 - Battleground 5: Whistleblower Protections<br>00:53:07 - Whistleblowing on Microsoft<br>00:54:43 - Andrej Karpathy &amp; Exercises in Empathy<br>01:05:57 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work</a></li><li><a href="https://www.susmangodfrey.com">Susman Godfrey LLP</a></li><li>Rich Sutton<ul><li><a href="http://incompleteideas.net/book/RLbook2020trimmed.pdf">Reinforcement Learning: An Introduction</a> (textbook)</li><li><a href="https://www.youtube.com/watch?v=NgHFMolXs3U">AI Succession</a> (presentation by Rich Sutton)</li><li><a href="https://arxiv.org/abs/2208.11173">The Alberta Plan for AI Research</a></li></ul></li><li>Moore's Law<ul><li><a href="https://www.computerhistory.org/collections/catalog/102770836">The Future of Integrated Electronics</a> (original paper)</li><li><a href="https://www.computerhistory.org/siliconengine/moores-law-predicts-the-future-of-integrated-circuits/">Computer History Museum's entry on Moore's Law</a></li></ul></li><li><a href="https://en.wikipedia.org/wiki/Stochastic_gradient_descent">Stochastic gradient descent (SGD)</a> on Wikipedia</li><li>OpenAI <em>Drama</em><ul><li><a href="https://maxread.substack.com/p/the-interested-normies-guide-to-openai">Max Read's Substack post</a></li><li>Zvi Mowshowitz's Substack series, in order of posting<ul><li><a href="https://thezvi.substack.com/p/openai-facts-from-a-weekend?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Facts from a Weekend</a></li><li><a href="https://thezvi.substack.com/p/openai-the-battle-of-the-board?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: The Battle of the Board</a></li><li><a href="https://thezvi.substack.com/p/openai-altman-returns?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Altman Returns</a></li><li><a href="https://thezvi.substack.com/p/openai-leaks-confirm-the-story?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Leaks Confirm the Story</a> ← best singular post in the series</li><li><a href="https://thezvi.substack.com/p/openai-the-board-expands?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: The Board Expands</a></li></ul></li><li><a href="https://openai.com/blog/openai-announces-new-members-to-board-of-directors">Official OpenAI announcement</a></li></ul></li><li><a href="https://en.wikipedia.org/wiki/Writers_Guild_of_America">WGA</a> on Wikipedia</li><li><a href="https://en.wikipedia.org/wiki/SAG-AFTRA">SAG-AFTRA</a> on Wikipedia</li><li>Tesla's False Advertising<ul><li><a href="https://www.latimes.com/business/story/2023-12-11/tesla-dmv-false-advertising-charges">Tesla's response to the DMV's false-advertising allegations: What took so long?</a></li><li><a href="https://www.caranddriver.com/news/a35785277/tesla-fsd-california-self-driving/">Tesla Tells California DMV that FSD Is Not Capable of Autonomous Driving</a></li><li><a href="https://www.caranddriver.com/features/columns/a35769759/carty-self-driving-cars/">What to Call Full Self-Driving When It Isn't Full Self-Driving?</a></li><li><a href="https://www.cnbc.com/2022/03/15/tesla-fired-employee-who-posted-fsd-beta-videos-as-ai-addict-on-youtube.html">Tesla fired an employee after he posted driverless tech reviews on YouTube</a></li><li><a href="https://www.tesla.com/support/autopilot">Tesla's page</a> on Autopilot and Full Self-Driving</li></ul></li><li>Recycling<ul><li><a href="https://www.longmontleader.com/regional-news/boulder-county-recycling-center-stockpiles-accurately-sorted-recyclable-materials-2385879">Boulder County Recycling Center Stockpiles Accurately Sorted Recyclable Materials</a></li><li><a href="https://boulderweekly.com/news/out-of-sight-out-of-mind/">Out of sight, out of mind</a></li><li><a href="https://ecocycle.org/content/uploads/2022/06/2024-Recycling-Guidelines-for-Boulder-County.pdf">Boulder Eco-Cycle Recycling Guidelines</a></li></ul></li><li><a href="https://arxiv.org/abs/2310.06009">Divide-and-Conquer Dynamics in AI-Driven Disempowerment</a></li><li>Microsoft Whistleblower<ul><li><a href="https://www.axios.com/2024/03/08/ai-whistleblowers-microsoft-copilot-designer-google-gemini">Whistleblowers call out AI's flaws</a></li><li><a href="https://www.linkedin.com/feed/update/urn:li:activity:7171135079702753280/">Shane's LinkedIn post</a><ul><li><a href="https://media.licdn.com/dms/document/media/D561FAQFfYnpLbIn2Xg/feedshare-document-pdf-analyzed/0/1709731400675?e=1711584000&amp;v=beta&amp;t=W3CLljNWJ8YCQSEuMCR9bt7UoIxsGo24Epj9GddfN8U">Letters sent by Jones</a></li></ul></li></ul></li><li><a href="https://twitter.com/karpathy/status/1757600075281547344">Karpathy announces departure from OpenAI</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of <a href="https://www.stakeout.ai">StakeOut.AI</a>, a non-profit focused on making AI go well <em>for humans</em>, along with Harry Luk and one other individual, whose name has been removed due to requirements of her current position.</p><p>In addition to the normal links, I wanted to include the links to the petitions that Dr. Park mentions during the podcast. Note that the nonprofit which began these petitions, StakeOut.AI, has been dissolved.<br><a href="https://www.change.org/p/right-ai-laws-to-right-our-future-support-artificial-intelligence-safety-regulations-now">Right AI Laws, to Right Our Future: Support Artificial Intelligence Safety Regulations Now</a><br><a href="https://www.change.org/p/is-deepfake-illegal-not-yet-ban-deepfakes-to-protect-your-family-demand-deepfake-laws">Is Deepfake Illegal? Not Yet! Ban Deepfakes to Protect Your Family &amp; Demand Deepfake Laws</a><br><a href="https://www.change.org/p/ban-superintelligence-stop-ai-driven-human-extinction-risk">Ban Superintelligence: Stop AI-Driven Human Extinction Risk </a></p><p>00:00:54 - Intro<br>00:02:34 - Battleground 1: Copyright<br>00:06:28 - Battleground 2: Moral Critique of AI Collaborationists<br>00:08:15 - Rich Sutton<br>00:20:41 - OpenAI <em>Drama</em><br>00:34:28 - Battleground 3: Contract Negotiations for AI Ban Clauses<br>00:37:57 - Tesla, Autopilot, and FSD<br>00:40:02 - Recycling<br>00:47:40 - Battleground 4: New Laws and Policies<br>00:50:00 - Battleground 5: Whistleblower Protections<br>00:53:07 - Whistleblowing on Microsoft<br>00:54:43 - Andrej Karpathy &amp; Exercises in Empathy<br>01:05:57 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work</a></li><li><a href="https://www.susmangodfrey.com">Susman Godfrey LLP</a></li><li>Rich Sutton<ul><li><a href="http://incompleteideas.net/book/RLbook2020trimmed.pdf">Reinforcement Learning: An Introduction</a> (textbook)</li><li><a href="https://www.youtube.com/watch?v=NgHFMolXs3U">AI Succession</a> (presentation by Rich Sutton)</li><li><a href="https://arxiv.org/abs/2208.11173">The Alberta Plan for AI Research</a></li></ul></li><li>Moore's Law<ul><li><a href="https://www.computerhistory.org/collections/catalog/102770836">The Future of Integrated Electronics</a> (original paper)</li><li><a href="https://www.computerhistory.org/siliconengine/moores-law-predicts-the-future-of-integrated-circuits/">Computer History Museum's entry on Moore's Law</a></li></ul></li><li><a href="https://en.wikipedia.org/wiki/Stochastic_gradient_descent">Stochastic gradient descent (SGD)</a> on Wikipedia</li><li>OpenAI <em>Drama</em><ul><li><a href="https://maxread.substack.com/p/the-interested-normies-guide-to-openai">Max Read's Substack post</a></li><li>Zvi Mowshowitz's Substack series, in order of posting<ul><li><a href="https://thezvi.substack.com/p/openai-facts-from-a-weekend?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Facts from a Weekend</a></li><li><a href="https://thezvi.substack.com/p/openai-the-battle-of-the-board?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: The Battle of the Board</a></li><li><a href="https://thezvi.substack.com/p/openai-altman-returns?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Altman Returns</a></li><li><a href="https://thezvi.substack.com/p/openai-leaks-confirm-the-story?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Leaks Confirm the Story</a> ← best singular post in the series</li><li><a href="https://thezvi.substack.com/p/openai-the-board-expands?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: The Board Expands</a></li></ul></li><li><a href="https://openai.com/blog/openai-announces-new-members-to-board-of-directors">Official OpenAI announcement</a></li></ul></li><li><a href="https://en.wikipedia.org/wiki/Writers_Guild_of_America">WGA</a> on Wikipedia</li><li><a href="https://en.wikipedia.org/wiki/SAG-AFTRA">SAG-AFTRA</a> on Wikipedia</li><li>Tesla's False Advertising<ul><li><a href="https://www.latimes.com/business/story/2023-12-11/tesla-dmv-false-advertising-charges">Tesla's response to the DMV's false-advertising allegations: What took so long?</a></li><li><a href="https://www.caranddriver.com/news/a35785277/tesla-fsd-california-self-driving/">Tesla Tells California DMV that FSD Is Not Capable of Autonomous Driving</a></li><li><a href="https://www.caranddriver.com/features/columns/a35769759/carty-self-driving-cars/">What to Call Full Self-Driving When It Isn't Full Self-Driving?</a></li><li><a href="https://www.cnbc.com/2022/03/15/tesla-fired-employee-who-posted-fsd-beta-videos-as-ai-addict-on-youtube.html">Tesla fired an employee after he posted driverless tech reviews on YouTube</a></li><li><a href="https://www.tesla.com/support/autopilot">Tesla's page</a> on Autopilot and Full Self-Driving</li></ul></li><li>Recycling<ul><li><a href="https://www.longmontleader.com/regional-news/boulder-county-recycling-center-stockpiles-accurately-sorted-recyclable-materials-2385879">Boulder County Recycling Center Stockpiles Accurately Sorted Recyclable Materials</a></li><li><a href="https://boulderweekly.com/news/out-of-sight-out-of-mind/">Out of sight, out of mind</a></li><li><a href="https://ecocycle.org/content/uploads/2022/06/2024-Recycling-Guidelines-for-Boulder-County.pdf">Boulder Eco-Cycle Recycling Guidelines</a></li></ul></li><li><a href="https://arxiv.org/abs/2310.06009">Divide-and-Conquer Dynamics in AI-Driven Disempowerment</a></li><li>Microsoft Whistleblower<ul><li><a href="https://www.axios.com/2024/03/08/ai-whistleblowers-microsoft-copilot-designer-google-gemini">Whistleblowers call out AI's flaws</a></li><li><a href="https://www.linkedin.com/feed/update/urn:li:activity:7171135079702753280/">Shane's LinkedIn post</a><ul><li><a href="https://media.licdn.com/dms/document/media/D561FAQFfYnpLbIn2Xg/feedshare-document-pdf-analyzed/0/1709731400675?e=1711584000&amp;v=beta&amp;t=W3CLljNWJ8YCQSEuMCR9bt7UoIxsGo24Epj9GddfN8U">Letters sent by Jones</a></li></ul></li></ul></li><li><a href="https://twitter.com/karpathy/status/1757600075281547344">Karpathy announces departure from OpenAI</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 18 Mar 2024 09:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/58dfc367/2d88b642.mp3" length="63715835" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>3983</itunes:duration>
      <itunes:summary>Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of StakeOut.AI, a non-profit focused on making AI go well for humans, along with Harry Luk and one other individual, whose name has been removed due to requirements of her current position.In addition to the normal links, I wanted to include the links to the petitions that Dr. Park mentions during the podcast. Note that the nonprofit which began these petitions, StakeOut.AI, has been dissolved.Right AI Laws, to Right Our Future: Support Artificial Intelligence Safety Regulations NowIs Deepfake Illegal? Not Yet! Ban Deepfakes to Protect Your Family &amp;amp; Demand Deepfake LawsBan Superintelligence: Stop AI-Driven Human Extinction Risk
00:00:54 - Intro00:02:34 - Battleground 1: Copyright00:06:28 - Battleground 2: Moral Critique of AI Collaborationists00:08:15 - Rich Sutton00:20:41 - OpenAI Drama00:34:28 - Battleground 3: Contract Negotiations for AI Ban Clauses00:37:57 - Tesla, Autopilot, and FSD00:40:02 - Recycling00:47:40 - Battleground 4: New Laws and Policies00:50:00 - Battleground 5: Whistleblower Protections00:53:07 - Whistleblowing on Microsoft00:54:43 - Andrej Karpathy &amp;amp; Exercises in Empathy01:05:57 - OutroLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.StakeOut.AIThe Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted WorkSusman Godfrey LLPRich SuttonReinforcement Learning: An Introduction (textbook)AI Succession (presentation by Rich Sutton)The Alberta Plan for AI Research Moore's LawThe Future of Integrated Electronics (original paper)Computer History Museum's entry on Moore's LawStochastic gradient descent (SGD) on WikipediaOpenAI DramaMax Read's Substack postZvi Mowshowitz's Substack series, in order of postingOpenAI: Facts from a WeekendOpenAI: The Battle of the BoardOpenAI: Altman ReturnsOpenAI: Leaks Confirm the Story ← best singular post in the seriesOpenAI: The Board ExpandsOfficial OpenAI announcementWGA on WikipediaSAG-AFTRA on WikipediaTesla's False AdvertisingTesla's response to the DMV's false-advertising allegations: What took so long?Tesla Tells California DMV that FSD Is Not Capable of Autonomous DrivingWhat to Call Full Self-Driving When It Isn't Full Self-Driving?Tesla fired an employee after he posted driverless tech reviews on YouTubeTesla's page on Autopilot and Full Self-DrivingRecyclingBoulder County Recycling Center Stockpiles Accurately Sorted Recyclable MaterialsOut of sight, out of mindBoulder Eco-Cycle Recycling GuidelinesDivide-and-Conquer Dynamics in AI-Driven DisempowermentMicrosoft WhistleblowerWhistleblowers call out AI's flawsShane's LinkedIn postLetters sent by JonesKarpathy announces departure from OpenAI</itunes:summary>
      <itunes:subtitle>Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of StakeOut.AI, a non-profit focused on making AI go well for humans, along with Harry Luk and one other in</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://scholar.harvard.edu/pspark" img="https://img.transistorcdn.com/uwPq3iaX1my6lU3bJjoP7cJ1SyqboZxZpJmB5Sw8bDc/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9mNjIz/ZmM0ZGJlMDU0NDZj/NDQ4NGVjNjE0YjFm/NDE1MC5qcGc.jpg">Dr. Peter S. Park</podcast:person>
    </item>
    <item>
      <title>MINISODE: Restructure Vol. 2</title>
      <itunes:episode>16</itunes:episode>
      <podcast:episode>16</podcast:episode>
      <itunes:title>MINISODE: Restructure Vol. 2</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">6dc7d6e0-46ac-49ce-a60f-fe7c86e3f648</guid>
      <link>https://kairos.fm/intoaisafety/e016</link>
      <description>
        <![CDATA[<p>UPDATE: Contrary to what I say in this episode, I won't be removing <em>any</em> episodes that are already published from the podcast RSS feed.</p><p>After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the podcast towards accessible content regarding "AI" instead of the show's original focus. I will still be releasing what I am calling research ride-along content to my <a href="https://www.patreon.com/IntoAISafety">Patreon</a>, but the show's feed will consist only of content that I aim to make as accessible as possible.</p><p>00:35 - TL;DL<br>01:12 - Advice from Pete<br>03:10 - My personal goal<br>05:39 - Reflection on refining my goal<br>09:08 - Looking forward (logistics</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>UPDATE: Contrary to what I say in this episode, I won't be removing <em>any</em> episodes that are already published from the podcast RSS feed.</p><p>After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the podcast towards accessible content regarding "AI" instead of the show's original focus. I will still be releasing what I am calling research ride-along content to my <a href="https://www.patreon.com/IntoAISafety">Patreon</a>, but the show's feed will consist only of content that I aim to make as accessible as possible.</p><p>00:35 - TL;DL<br>01:12 - Advice from Pete<br>03:10 - My personal goal<br>05:39 - Reflection on refining my goal<br>09:08 - Looking forward (logistics</p>]]>
      </content:encoded>
      <pubDate>Mon, 11 Mar 2024 09:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/fa6df270/5048eeef.mp3" length="12611516" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>789</itunes:duration>
      <itunes:summary>UPDATE: Contrary to what I say in this episode, I won't be removing any episodes that are already published from the podcast RSS feed.
After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the podcast towards accessible content regarding "AI" instead of the show's original focus. I will still be releasing what I am calling research ride-along content to my Patreon, but the show's feed will consist only of content that I aim to make as accessible as possible.
00:35 - TL;DL01:12 - Advice from Pete03:10 - My personal goal05:39 - Reflection on refining my goal09:08 - Looking forward (logistics</itunes:summary>
      <itunes:subtitle>UPDATE: Contrary to what I say in this episode, I won't be removing any episodes that are already published from the podcast RSS feed.
After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the po</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (1)</title>
      <itunes:episode>15</itunes:episode>
      <podcast:episode>15</podcast:episode>
      <itunes:title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (1)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">cd0206b8-6690-44c7-97e3-240c2b082096</guid>
      <link>https://kairos.fm/intoaisafety/e015</link>
      <description>
        <![CDATA[<p>Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded <a href="https://www.stakeout.ai/">⁠StakeOut.AI</a>, a non-profit focused on making AI go well <em>for humans</em>.</p><p>00:54 - Intro<br>03:15 - Dr. Park, x-risk, and AGI<br>08:55 - StakeOut.AI<br>12:05 - Governance scorecard<br>19:34 - Hollywood webinar<br>22:02 - Regulations.gov comments<br>23:48 - Open letters <br>26:15 - EU AI Act<br>35:07 - Effective accelerationism<br>40:50 - Divide and conquer dynamics<br>45:40 - AI "art"<br>53:09 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://futureoflife.org/wp-content/uploads/2023/11/FLI_Governance_Scorecard_and_Framework.pdf">AI Governance Scorecard</a> (go to Pg. 3)</li><li><a href="https://pauseai.info">Pause AI</a></li><li><a href="https://www.regulations.gov">Regulations.gov</a><br> <ul><li><a href="https://www.regulations.gov/comment/COLC-2023-0006-10077">USCO StakeOut.AI Comment</a></li></ul></li><li> <ul><li><a href="https://www.regulations.gov/comment/OMB-2023-0020-0170">OMB StakeOut.AI Comment</a></li></ul></li><li> </li><li><a href="https://aitreaty.org">AI Treaty open letter</a></li><li><a href="https://taisc.org">TAISC</a></li><li><a href="https://crfm.stanford.edu/2023/03/13/alpaca.html">Alpaca: A Strong, Replicable Instruction-Following Model</a></li><li>References on EU AI Act and Cedric O<br> <ul><li><a href="https://twitter.com/cedric_o/status/1728724005459235052">Tweet from Cedric O</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/eu-policymakers-enter-the-last-mile-for-artificial-intelligence-rulebook/">EU policymakers enter the last mile for Artificial Intelligence rulebook</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/ai-act-eu-parliaments-legal-office-gives-damning-opinion-on-high-risk-classification-filters/">AI Act: EU Parliament’s legal office gives damning opinion on high-risk classification ‘filters’</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/eus-ai-act-negotiations-hit-the-brakes-over-foundation-models/">EU’s AI Act negotiations hit the brakes over foundation models</a></li></ul></li><li> <ul><li><a href="https://www.foundation-models.eu">The EU AI Act needs Foundation Model Regulation</a></li></ul></li><li> <ul><li><a href="https://verfassungsblog.de/bigtechs-efforts-to-derail-the-ai-act/">BigTech’s Efforts to Derail the AI Act</a></li></ul></li><li> </li><li><a href="https://demos.co.uk/research/open-sourcing-the-ai-revolution-framing-the-debate-on-open-source-artificial-intelligence-and-regulation/">Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation</a></li><li><a href="https://arxiv.org/abs/2310.06009">Divide-and-Conquer Dynamics in AI-Driven Disempowerment</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded <a href="https://www.stakeout.ai/">⁠StakeOut.AI</a>, a non-profit focused on making AI go well <em>for humans</em>.</p><p>00:54 - Intro<br>03:15 - Dr. Park, x-risk, and AGI<br>08:55 - StakeOut.AI<br>12:05 - Governance scorecard<br>19:34 - Hollywood webinar<br>22:02 - Regulations.gov comments<br>23:48 - Open letters <br>26:15 - EU AI Act<br>35:07 - Effective accelerationism<br>40:50 - Divide and conquer dynamics<br>45:40 - AI "art"<br>53:09 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://futureoflife.org/wp-content/uploads/2023/11/FLI_Governance_Scorecard_and_Framework.pdf">AI Governance Scorecard</a> (go to Pg. 3)</li><li><a href="https://pauseai.info">Pause AI</a></li><li><a href="https://www.regulations.gov">Regulations.gov</a><br> <ul><li><a href="https://www.regulations.gov/comment/COLC-2023-0006-10077">USCO StakeOut.AI Comment</a></li></ul></li><li> <ul><li><a href="https://www.regulations.gov/comment/OMB-2023-0020-0170">OMB StakeOut.AI Comment</a></li></ul></li><li> </li><li><a href="https://aitreaty.org">AI Treaty open letter</a></li><li><a href="https://taisc.org">TAISC</a></li><li><a href="https://crfm.stanford.edu/2023/03/13/alpaca.html">Alpaca: A Strong, Replicable Instruction-Following Model</a></li><li>References on EU AI Act and Cedric O<br> <ul><li><a href="https://twitter.com/cedric_o/status/1728724005459235052">Tweet from Cedric O</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/eu-policymakers-enter-the-last-mile-for-artificial-intelligence-rulebook/">EU policymakers enter the last mile for Artificial Intelligence rulebook</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/ai-act-eu-parliaments-legal-office-gives-damning-opinion-on-high-risk-classification-filters/">AI Act: EU Parliament’s legal office gives damning opinion on high-risk classification ‘filters’</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/eus-ai-act-negotiations-hit-the-brakes-over-foundation-models/">EU’s AI Act negotiations hit the brakes over foundation models</a></li></ul></li><li> <ul><li><a href="https://www.foundation-models.eu">The EU AI Act needs Foundation Model Regulation</a></li></ul></li><li> <ul><li><a href="https://verfassungsblog.de/bigtechs-efforts-to-derail-the-ai-act/">BigTech’s Efforts to Derail the AI Act</a></li></ul></li><li> </li><li><a href="https://demos.co.uk/research/open-sourcing-the-ai-revolution-framing-the-debate-on-open-source-artificial-intelligence-and-regulation/">Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation</a></li><li><a href="https://arxiv.org/abs/2310.06009">Divide-and-Conquer Dynamics in AI-Driven Disempowerment</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 04 Mar 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/81a6fb5c/cb1fde4e.mp3" length="52004201" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>3251</itunes:duration>
      <itunes:summary>Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded 
⁠StakeOut.AI, a non-profit focused on making AI go well for humans.
00:54 - Intro03:15 - Dr. Park, x-risk, and AGI08:55 - StakeOut.AI12:05 - Governance scorecard19:34 - Hollywood webinar22:02 - Regulations.gov comments23:48 - Open letters 26:15 - EU AI Act35:07 - Effective accelerationism40:50 - Divide and conquer dynamics45:40 - AI "art"53:09 - Outro
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.

 StakeOut.AI
 AI Governance Scorecard (go to Pg. 3)
 Pause AI
 Regulations.gov
  USCO StakeOut.AI Comment
  OMB StakeOut.AI Comment


  AI Treaty open letter
  TAISC
  Alpaca: A Strong, Replicable Instruction-Following Model
  References on EU AI Act and Cedric O
  Tweet from Cedric O
  EU policymakers enter the last mile for Artificial Intelligence rulebook
  AI Act: EU Parliament’s legal office gives damning opinion on high-risk classification ‘filters’
  EU’s AI Act negotiations hit the brakes over foundation models
  The EU AI Act needs Foundation Model Regulation
  BigTech’s Efforts to Derail the AI Act


  Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation
  Divide-and-Conquer Dynamics in AI-Driven Disempowerment</itunes:summary>
      <itunes:subtitle>Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded 
⁠StakeOut.AI, a non-profit focused on making AI go well for humans.
00:54 - Intro03:15</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://scholar.harvard.edu/pspark" img="https://img.transistorcdn.com/uwPq3iaX1my6lU3bJjoP7cJ1SyqboZxZpJmB5Sw8bDc/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9mNjIz/ZmM0ZGJlMDU0NDZj/NDQ4NGVjNjE0YjFm/NDE1MC5qcGc.jpg">Dr. Peter S. Park</podcast:person>
    </item>
    <item>
      <title>MINISODE: "LLMs, a Survey"</title>
      <itunes:episode>14</itunes:episode>
      <podcast:episode>14</podcast:episode>
      <itunes:title>MINISODE: "LLMs, a Survey"</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">2d7e4db8-5798-4148-9e19-53a5143c4919</guid>
      <link>https://kairos.fm/intoaisafety/e014</link>
      <description>
        <![CDATA[<p>Take a trip with me through the paper <a href="https://arxiv.org/abs/2402.06196">Large Language Models, A Survey</a>, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety <a href="https://into-ai-safety.github.io">podcast website</a>.</p><p>00:36 - Intro and authors<br>01:50 - My takes and paper structure<br>04:40 - Getting to LLMs<br>07:27 - Defining LLMs &amp; emergence<br>12:12 - Overview of PLMs<br>15:00 - How LLMs are built<br>18:52 - Limitations if LLMs<br>23:06 - Uses of LLMs<br>25:16 - Evaluations and Benchmarks<br>28:11 - Challenges and future directions<br>29:21 - Recap &amp; outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://arxiv.org/abs/2402.06196">Large Language Models, A Survey</a></li><li><a href="https://www.linkedin.com/posts/meysam-ac_i-am-delighted-to-share-that-our-most-recent-activity-7162768857827377152-wiLu/?utm_source=share&amp;utm_medium=member_desktop">Meysam's LinkedIn Post</a></li><li>Claude E. Shannon<ul><li><a href="https://dspace.mit.edu/handle/1721.1/11173">A symbolic analysis of relay and switching circuits</a> (Master's Thesis)</li><li><a href="https://ieeexplore.ieee.org/document/6769090">Communication theory of secrecy systems</a></li><li><a href="https://ieeexplore.ieee.org/document/6773024">A mathematical theory of communication</a></li><li><a href="https://ieeexplore.ieee.org/document/6773263">Prediction and entropy of printed English</a></li></ul></li><li><a href="https://bounded-regret.ghost.io/future-ml-systems-will-be-qualitatively-different/">Future ML Systems Will Be Qualitatively Different</a></li><li><a href="https://www.science.org/doi/10.1126/science.177.4047.393?ref=bounded-regret.ghost.io">More Is Different</a></li><li><a href="https://arxiv.org/abs/2401.05566">Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training</a></li><li><a href="https://arxiv.org/abs/2304.15004">Are Emergent Abilities of Large Language Models a Mirage?</a></li><li><a href="https://arxiv.org/abs/2309.01809">Are Emergent Abilities of Large Language Models just In-Context Learning?</a></li><li><a href="https://arxiv.org/abs/1706.03762">Attention is all you need</a></li><li><a href="https://arxiv.org/abs/2305.18290">Direct Preference Optimization: Your Language Model is Secretly a Reward Model</a></li><li><a href="https://arxiv.org/abs/2402.01306">KTO: Model Alignment as Prospect Theoretic Optimization</a></li><li><a href="https://www2.stat.duke.edu/~scs/Courses/Stat376/Papers/TemperAnneal/KirkpatrickAnnealScience1983.pdf">Optimization by Simulated Annealing</a></li><li><a href="https://openai.com/blog/memory-and-new-controls-for-chatgpt">Memory and new controls for ChatGPT</a></li><li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4515540/">Hallucinations and related concepts—their conceptual background</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Take a trip with me through the paper <a href="https://arxiv.org/abs/2402.06196">Large Language Models, A Survey</a>, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety <a href="https://into-ai-safety.github.io">podcast website</a>.</p><p>00:36 - Intro and authors<br>01:50 - My takes and paper structure<br>04:40 - Getting to LLMs<br>07:27 - Defining LLMs &amp; emergence<br>12:12 - Overview of PLMs<br>15:00 - How LLMs are built<br>18:52 - Limitations if LLMs<br>23:06 - Uses of LLMs<br>25:16 - Evaluations and Benchmarks<br>28:11 - Challenges and future directions<br>29:21 - Recap &amp; outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://arxiv.org/abs/2402.06196">Large Language Models, A Survey</a></li><li><a href="https://www.linkedin.com/posts/meysam-ac_i-am-delighted-to-share-that-our-most-recent-activity-7162768857827377152-wiLu/?utm_source=share&amp;utm_medium=member_desktop">Meysam's LinkedIn Post</a></li><li>Claude E. Shannon<ul><li><a href="https://dspace.mit.edu/handle/1721.1/11173">A symbolic analysis of relay and switching circuits</a> (Master's Thesis)</li><li><a href="https://ieeexplore.ieee.org/document/6769090">Communication theory of secrecy systems</a></li><li><a href="https://ieeexplore.ieee.org/document/6773024">A mathematical theory of communication</a></li><li><a href="https://ieeexplore.ieee.org/document/6773263">Prediction and entropy of printed English</a></li></ul></li><li><a href="https://bounded-regret.ghost.io/future-ml-systems-will-be-qualitatively-different/">Future ML Systems Will Be Qualitatively Different</a></li><li><a href="https://www.science.org/doi/10.1126/science.177.4047.393?ref=bounded-regret.ghost.io">More Is Different</a></li><li><a href="https://arxiv.org/abs/2401.05566">Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training</a></li><li><a href="https://arxiv.org/abs/2304.15004">Are Emergent Abilities of Large Language Models a Mirage?</a></li><li><a href="https://arxiv.org/abs/2309.01809">Are Emergent Abilities of Large Language Models just In-Context Learning?</a></li><li><a href="https://arxiv.org/abs/1706.03762">Attention is all you need</a></li><li><a href="https://arxiv.org/abs/2305.18290">Direct Preference Optimization: Your Language Model is Secretly a Reward Model</a></li><li><a href="https://arxiv.org/abs/2402.01306">KTO: Model Alignment as Prospect Theoretic Optimization</a></li><li><a href="https://www2.stat.duke.edu/~scs/Courses/Stat376/Papers/TemperAnneal/KirkpatrickAnnealScience1983.pdf">Optimization by Simulated Annealing</a></li><li><a href="https://openai.com/blog/memory-and-new-controls-for-chatgpt">Memory and new controls for ChatGPT</a></li><li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4515540/">Hallucinations and related concepts—their conceptual background</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 26 Feb 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/24982632/d6945322.mp3" length="29667591" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>1855</itunes:duration>
      <itunes:summary>Take a trip with me through the paper Large Language Models, A Survey, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety podcast website.
00:36 - Intro and authors01:50 - My takes and paper structure04:40 - Getting to LLMs07:27 - Defining LLMs &amp;amp; emergence12:12 - Overview of PLMs15:00 - How LLMs are built18:52 - Limitations if LLMs23:06 - Uses of LLMs25:16 - Evaluations and Benchmarks28:11 - Challenges and future directions29:21 - Recap &amp;amp; outro
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.Large Language Models, A SurveyMeysam's LinkedIn PostClaude E. ShannonA symbolic analysis of relay and switching circuits (Master's Thesis)Communication theory of secrecy systemsA mathematical theory of communicationPrediction and entropy of printed EnglishFuture ML Systems Will Be Qualitatively DifferentMore Is DifferentSleeper Agents: Training Deceptive LLMs that Persist Through Safety TrainingAre Emergent Abilities of Large Language Models a Mirage?Are Emergent Abilities of Large Language Models just In-Context Learning?Attention is all you needDirect Preference Optimization: Your Language Model is Secretly a Reward ModelKTO: Model Alignment as Prospect Theoretic OptimizationOptimization by Simulated AnnealingMemory and new controls for ChatGPTHallucinations and related concepts—their conceptual background</itunes:summary>
      <itunes:subtitle>Take a trip with me through the paper Large Language Models, A Survey, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety podcast website.
00:36 - Intro and authors01:50 - My take</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>FEEDBACK: Applying for Funding w/ Esben Kran</title>
      <itunes:episode>13</itunes:episode>
      <podcast:episode>13</podcast:episode>
      <itunes:title>FEEDBACK: Applying for Funding w/ Esben Kran</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">9c9288c4-d48b-48fa-aec6-b11cc5803619</guid>
      <link>https://kairos.fm/intoaisafety/e013</link>
      <description>
        <![CDATA[<p>Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when applying for funding in the future.</p><p>Head over to Apart Research's <a href="https://apartresearch.com">website</a> to check out their work, or the Alignment Jam <a href="https://alignmentjam.com">website</a> for information on upcoming hackathons.</p><p>A doc-capsule of the application at the time of this recording can be found <a href="https://docs.google.com/document/d/1ofk5VLvNeJytd5Rpx18zui-JfV900DZv5nZShZcrgfY/edit?usp=sharing">at this link</a>.</p><p>01:38 - Interview starts<br>05:41 - Proposal<br>11:00 - Personal statement<br>14:00 - Budget<br>21:12 - CV<br>22:45 - Application questions<br>34:06 - Funding questions<br>44:25 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.openphilanthropy.org/research/ai-governance-talent-profiles-wed-like-to-see/">AI governance talent profiles we’d like to see</a></li><li><a href="https://alignmentjam.com/jam/governance">The AI Governance Research Sprint</a></li><li><a href="https://www.openphilanthropy.org/research/reasoning-transparency/">Reasoning Transparency</a></li><li>Places to look for funding<ul><li>Open Philanthropy's <a href="https://www.openphilanthropy.org/career-development-and-transition-funding/">Career development and transition funding</a></li><li><a href="https://funds.effectivealtruism.org/funds/far-future">Long-Term Future Fund</a></li><li><a href="https://manifund.org">Manifund</a></li></ul></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when applying for funding in the future.</p><p>Head over to Apart Research's <a href="https://apartresearch.com">website</a> to check out their work, or the Alignment Jam <a href="https://alignmentjam.com">website</a> for information on upcoming hackathons.</p><p>A doc-capsule of the application at the time of this recording can be found <a href="https://docs.google.com/document/d/1ofk5VLvNeJytd5Rpx18zui-JfV900DZv5nZShZcrgfY/edit?usp=sharing">at this link</a>.</p><p>01:38 - Interview starts<br>05:41 - Proposal<br>11:00 - Personal statement<br>14:00 - Budget<br>21:12 - CV<br>22:45 - Application questions<br>34:06 - Funding questions<br>44:25 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.openphilanthropy.org/research/ai-governance-talent-profiles-wed-like-to-see/">AI governance talent profiles we’d like to see</a></li><li><a href="https://alignmentjam.com/jam/governance">The AI Governance Research Sprint</a></li><li><a href="https://www.openphilanthropy.org/research/reasoning-transparency/">Reasoning Transparency</a></li><li>Places to look for funding<ul><li>Open Philanthropy's <a href="https://www.openphilanthropy.org/career-development-and-transition-funding/">Career development and transition funding</a></li><li><a href="https://funds.effectivealtruism.org/funds/far-future">Long-Term Future Fund</a></li><li><a href="https://manifund.org">Manifund</a></li></ul></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 19 Feb 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/aa6ca492/29c49e43.mp3" length="43395914" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>2713</itunes:duration>
      <itunes:summary>Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when applying for funding in the future.Head over to Apart Research's website to check out their work, or the Alignment Jam website for information on upcoming hackathons.A doc-capsule of the application at the time of this recording can be found at this link.01:38 - Interview starts05:41 - Proposal11:00 - Personal statement14:00 - Budget21:12 - CV22:45 - Application questions34:06 - Funding questions44:25 - OutroLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.AI governance talent profiles we’d like to seeThe AI Governance Research SprintReasoning TransparencyPlaces to look for fundingOpen Philanthropy's Career development and transition fundingLong-Term Future FundManifund</itunes:summary>
      <itunes:subtitle>Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when </itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Reading a Research Paper</title>
      <itunes:episode>12</itunes:episode>
      <podcast:episode>12</podcast:episode>
      <itunes:title>MINISODE: Reading a Research Paper</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">37fd4eb5-ba2f-419e-aa24-9a1425a69051</guid>
      <link>https://kairos.fm/intoaisafety/e012</link>
      <description>
        <![CDATA[<p>Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice.</p><p>00:56 - Anatomy of a paper<br>02:38 - Most common advice<br>05:24 - Reading sparsity and path<br>07:30 - Notes and motivation</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7392212/">Ten simple rules for reading a scientific paper</a></li><li>Best sources I found<ul><li><a href="https://subjectguides.york.ac.uk/critical/articles">Let's get critical: Reading academic articles</a></li><li><a href="https://www.scientifica.uk.com/neurowire/gradhacks-a-guide-to-reading-research-papers">#GradHacks: A guide to reading research papers</a></li><li><a href="https://www.lib.purdue.edu/sites/default/files/libraries/engr/Tutorials/Newest%20Scientific%20Paper.pdf">How to read a scientific paper</a> (presentation)</li></ul></li><li>Some more sources<ul><li><a href="https://www.owlnet.rice.edu/~cainproj/courses/HowToReadSciArticle.pdf">How to read a scientific article</a></li><li><a href="https://www.eecs.harvard.edu/~michaelm/postscripts/ReadPaper.pdf">How to read a research paper</a></li><li><a href="https://resources.nu.edu/researchprocess/readingscientificarticle">Reading a scientific article</a></li></ul></li><li> </li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice.</p><p>00:56 - Anatomy of a paper<br>02:38 - Most common advice<br>05:24 - Reading sparsity and path<br>07:30 - Notes and motivation</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7392212/">Ten simple rules for reading a scientific paper</a></li><li>Best sources I found<ul><li><a href="https://subjectguides.york.ac.uk/critical/articles">Let's get critical: Reading academic articles</a></li><li><a href="https://www.scientifica.uk.com/neurowire/gradhacks-a-guide-to-reading-research-papers">#GradHacks: A guide to reading research papers</a></li><li><a href="https://www.lib.purdue.edu/sites/default/files/libraries/engr/Tutorials/Newest%20Scientific%20Paper.pdf">How to read a scientific paper</a> (presentation)</li></ul></li><li>Some more sources<ul><li><a href="https://www.owlnet.rice.edu/~cainproj/courses/HowToReadSciArticle.pdf">How to read a scientific article</a></li><li><a href="https://www.eecs.harvard.edu/~michaelm/postscripts/ReadPaper.pdf">How to read a research paper</a></li><li><a href="https://resources.nu.edu/researchprocess/readingscientificarticle">Reading a scientific article</a></li></ul></li><li> </li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 12 Feb 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/36d075b9/8a9a30fe.mp3" length="9027946" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>565</itunes:duration>
      <itunes:summary>Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice.00:56 - Anatomy of a paper02:38 - Most common advice05:24 - Reading sparsity and path07:30 - Notes and motivationLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.Ten simple rules for reading a scientific paperBest sources I foundLet's get critical: Reading academic articles#GradHacks: A guide to reading research papersHow to read a scientific paper (presentation)Some more sourcesHow to read a scientific articleHow to read a research paperReading a scientific article</itunes:summary>
      <itunes:subtitle>Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice.00:56 - Anatomy of a paper02:38 - Most comm</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>HACKATHON: Evals November 2023 (2)</title>
      <itunes:episode>11</itunes:episode>
      <podcast:episode>11</podcast:episode>
      <itunes:title>HACKATHON: Evals November 2023 (2)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">5e003e1c-4e32-487d-905f-9672cc241a26</guid>
      <link>https://kairos.fm/intoaisafety/e011</link>
      <description>
        <![CDATA[<p>Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation.</p><p>Check out Stellaric's <a href="https://stellaric.pw/">website</a>, or follow them on <a href="https://twitter.com/stellaricpw">Twitter</a>.</p><p>01:53 - Meeting starts<br>05:05 - Pitch: extension of locked models<br>23:23 - Pitch: retroactive holdout datasets<br>34:04 - Preliminary results<br>37:44 - Next steps<br>42:55 - Recap</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://github.com/LRudL/evalugator">Evalugator</a> library</li><li><a href="https://www.alignmentforum.org/posts/rZs6ddqNnW8LXuJqA/password-locked-models-a-stress-case-for-capabilities">Password Locked Model</a> blogpost</li><li><a href="https://arxiv.org/abs/2109.07958">TruthfulQA: Measuring How Models Mimic Human Falsehoods</a></li><li><a href="https://aclanthology.org/P02-1040.pdf">BLEU: a Method for Automatic Evaluation of Machine Translation</a></li><li><a href="https://arxiv.org/abs/1905.10044">BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions</a></li><li><a href="https://arxiv.org/abs/2310.16789">Detecting Pretraining Data from Large Language Models</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation.</p><p>Check out Stellaric's <a href="https://stellaric.pw/">website</a>, or follow them on <a href="https://twitter.com/stellaricpw">Twitter</a>.</p><p>01:53 - Meeting starts<br>05:05 - Pitch: extension of locked models<br>23:23 - Pitch: retroactive holdout datasets<br>34:04 - Preliminary results<br>37:44 - Next steps<br>42:55 - Recap</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://github.com/LRudL/evalugator">Evalugator</a> library</li><li><a href="https://www.alignmentforum.org/posts/rZs6ddqNnW8LXuJqA/password-locked-models-a-stress-case-for-capabilities">Password Locked Model</a> blogpost</li><li><a href="https://arxiv.org/abs/2109.07958">TruthfulQA: Measuring How Models Mimic Human Falsehoods</a></li><li><a href="https://aclanthology.org/P02-1040.pdf">BLEU: a Method for Automatic Evaluation of Machine Translation</a></li><li><a href="https://arxiv.org/abs/1905.10044">BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions</a></li><li><a href="https://arxiv.org/abs/2310.16789">Detecting Pretraining Data from Large Language Models</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 05 Feb 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/b1493509/6a5ceac9.mp3" length="46699824" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>2919</itunes:duration>
      <itunes:summary>Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation.Check out Stellaric's website, or follow them on Twitter.01:53 - Meeting starts05:05 - Pitch: extension of locked models23:23 - Pitch: retroactive holdout datasets34:04 - Preliminary results37:44 - Next steps42:55 - RecapLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.Evalugator libraryPassword Locked Model blogpostTruthfulQA: Measuring How Models Mimic Human FalsehoodsBLEU: a Method for Automatic Evaluation of Machine TranslationBoolQ: Exploring the Surprising Difficulty of Natural Yes/No QuestionsDetecting Pretraining Data from Large Language Models</itunes:summary>
      <itunes:subtitle>Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation.Check out Stellaric's website, or follow them on T</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Portfolios</title>
      <itunes:episode>10</itunes:episode>
      <podcast:episode>10</podcast:episode>
      <itunes:title>MINISODE: Portfolios</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">d42a5525-a155-4891-9cd1-98b1af663463</guid>
      <link>https://kairos.fm/intoaisafety/e010</link>
      <description>
        <![CDATA[<p>I provide my thoughts and recommendations regarding personal professional portfolios.</p><p>00:35 - Intro to portfolios<br>01:42 - Modern portfolios<br>02:27 - What to include<br>04:38 - Importance of visual<br>05:50 - The "About" page<br>06:25 - Tools<br>08:12 - Future of "Minisodes"</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://jime.open.ac.uk/articles/10.5334/jime.574">From Portafoglio to Eportfolio: The Evolution of Portfolio in Higher Education</a></li><li><a href="https://www.gimp.org">GIMP</a></li><li><a href="https://alternativeto.net">AlternativeTo</a></li><li><a href="https://jekyllrb.com">Jekyll</a></li><li><a href="https://pages.github.com">GitHub Pages</a></li><li><a href="https://mmistakes.github.io/minimal-mistakes/">Minimal Mistakes</a></li><li><a href="https://jacob-haimes.github.io">My portfolio</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>I provide my thoughts and recommendations regarding personal professional portfolios.</p><p>00:35 - Intro to portfolios<br>01:42 - Modern portfolios<br>02:27 - What to include<br>04:38 - Importance of visual<br>05:50 - The "About" page<br>06:25 - Tools<br>08:12 - Future of "Minisodes"</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://jime.open.ac.uk/articles/10.5334/jime.574">From Portafoglio to Eportfolio: The Evolution of Portfolio in Higher Education</a></li><li><a href="https://www.gimp.org">GIMP</a></li><li><a href="https://alternativeto.net">AlternativeTo</a></li><li><a href="https://jekyllrb.com">Jekyll</a></li><li><a href="https://pages.github.com">GitHub Pages</a></li><li><a href="https://mmistakes.github.io/minimal-mistakes/">Minimal Mistakes</a></li><li><a href="https://jacob-haimes.github.io">My portfolio</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 29 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/0258007f/e8a8e309.mp3" length="9251905" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>579</itunes:duration>
      <itunes:summary>I provide my thoughts and recommendations regarding personal professional portfolios.
00:35 - Intro to portfolios01:42 - Modern portfolios02:27 - What to include04:38 - Importance of visual05:50 - The "About" page06:25 - Tools08:12 - Future of "Minisodes"
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.From Portafoglio to Eportfolio: The Evolution of Portfolio in Higher EducationGIMPAlternativeToJekyllGitHub PagesMinimal MistakesMy portfolio</itunes:summary>
      <itunes:subtitle>I provide my thoughts and recommendations regarding personal professional portfolios.
00:35 - Intro to portfolios01:42 - Modern portfolios02:27 - What to include04:38 - Importance of visual05:50 - The "About" page06:25 - Tools08:12 - Future of "Minisodes</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>INTERVIEW: Polysemanticity w/ Dr. Darryl Wright</title>
      <itunes:episode>9</itunes:episode>
      <podcast:episode>9</podcast:episode>
      <itunes:title>INTERVIEW: Polysemanticity w/ Dr. Darryl Wright</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">f4ca1b13-2c6d-4c18-a458-9131448382b3</guid>
      <link>https://kairos.fm/intoaisafety/e009</link>
      <description>
        <![CDATA[<p>Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks.</p><p>Check out a <a href="https://into-ai-safety.github.io/episode/interview/episode-8/">diagram</a> of the decoder task used for our research!</p><p>01:46 - Interview begins<br>02:14 - Supernovae classification<br>08:58 - Penalizing polysemanticity<br>20:58 - Our "toy model"<br>30:06 - Task description<br>32:47 - Addressing hurdles<br>39:20 - Lessons learned</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.zooniverse.org">Zooniverse</a></li><li><a href="https://bluedot.org">BlueDot Impact</a></li><li><a href="https://www.aisafetysupport.org">AI Safety Support</a></li><li><a href="https://distill.pub/2020/circuits/zoom-in/">Zoom In: An Introduction to Circuits</a></li><li><a href="https://paperswithcode.com/dataset/mnist">MNIST</a> dataset on PapersWithCode</li><li><a href="https://arxiv.org/abs/2103.03386">Clusterability in Neural Networks</a></li><li><a href="https://www.cs.toronto.edu/~kriz/cifar.html">CIFAR-10</a> dataset</li><li><a href="https://www.effectivealtruism.org/ea-global">Effective Altruism Global</a></li><li><a href="https://openai.com/research/clip">CLIP</a> (blog post)</li><li><a href="https://funds.effectivealtruism.org/funds/far-future">Long Term Future Fund</a></li><li><a href="https://arxiv.org/abs/2211.09169">Engineering Monosemanticity in Toy Models</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks.</p><p>Check out a <a href="https://into-ai-safety.github.io/episode/interview/episode-8/">diagram</a> of the decoder task used for our research!</p><p>01:46 - Interview begins<br>02:14 - Supernovae classification<br>08:58 - Penalizing polysemanticity<br>20:58 - Our "toy model"<br>30:06 - Task description<br>32:47 - Addressing hurdles<br>39:20 - Lessons learned</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.zooniverse.org">Zooniverse</a></li><li><a href="https://bluedot.org">BlueDot Impact</a></li><li><a href="https://www.aisafetysupport.org">AI Safety Support</a></li><li><a href="https://distill.pub/2020/circuits/zoom-in/">Zoom In: An Introduction to Circuits</a></li><li><a href="https://paperswithcode.com/dataset/mnist">MNIST</a> dataset on PapersWithCode</li><li><a href="https://arxiv.org/abs/2103.03386">Clusterability in Neural Networks</a></li><li><a href="https://www.cs.toronto.edu/~kriz/cifar.html">CIFAR-10</a> dataset</li><li><a href="https://www.effectivealtruism.org/ea-global">Effective Altruism Global</a></li><li><a href="https://openai.com/research/clip">CLIP</a> (blog post)</li><li><a href="https://funds.effectivealtruism.org/funds/far-future">Long Term Future Fund</a></li><li><a href="https://arxiv.org/abs/2211.09169">Engineering Monosemanticity in Toy Models</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 22 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/0afbbf3c/a966e060.mp3" length="43331933" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>2709</itunes:duration>
      <itunes:summary>Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks.
Check out a diagram of the decoder task used for our research!
01:46 - Interview begins02:14 - Supernovae classification08:58 - Penalizing polysemanticity20:58 - Our "toy model"30:06 - Task description32:47 - Addressing hurdles39:20 - Lessons learned
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.

 Zooniverse
 BlueDot Impact
 AI Safety Support
 Zoom In: An Introduction to Circuits
  MNIST dataset on PapersWithCode
  Clusterability in Neural Networks
  CIFAR-10 dataset
  Effective Altruism Global
  CLIP (blog post)
  Long Term Future Fund
  Engineering Monosemanticity in Toy Models</itunes:summary>
      <itunes:subtitle>Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks.
Check out a diagram of the decoder task</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Starting a Podcast</title>
      <itunes:episode>8</itunes:episode>
      <podcast:episode>8</podcast:episode>
      <itunes:title>MINISODE: Starting a Podcast</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">24d72baa-2a3e-4ae1-8d64-2af3b7cb34bb</guid>
      <link>https://kairos.fm/intoaisafety/e008</link>
      <description>
        <![CDATA[<p>A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://podcasters.spotify.com/">Spotify for Podcasters</a></li><li><a href="https://into-ai-safety.github.io">Into AI Safety podcast website</a></li><li><a href="https://www.effectivealtruism.org/ea-global">Effective Altruism Global</a></li><li><a href="https://obsproject.com">Open Broadcaster Software (OBS)</a></li><li><a href="https://craig.chat">Craig</a></li><li><a href="https://riverside.fm">Riverside</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://podcasters.spotify.com/">Spotify for Podcasters</a></li><li><a href="https://into-ai-safety.github.io">Into AI Safety podcast website</a></li><li><a href="https://www.effectivealtruism.org/ea-global">Effective Altruism Global</a></li><li><a href="https://obsproject.com">Open Broadcaster Software (OBS)</a></li><li><a href="https://craig.chat">Craig</a></li><li><a href="https://riverside.fm">Riverside</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 15 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/4823b96d/02a05da4.mp3" length="10148026" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>635</itunes:duration>
      <itunes:summary>A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar.Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.LessWrongSpotify for PodcastersInto AI Safety podcast websiteEffective Altruism GlobalOpen Broadcaster Software (OBS)CraigRiverside</itunes:summary>
      <itunes:subtitle>A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar.Links to all articles/papers which are mentioned throughout the episode can be found below</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>HACKATHON: Evals November 2023 (1)</title>
      <itunes:episode>7</itunes:episode>
      <podcast:episode>7</podcast:episode>
      <itunes:title>HACKATHON: Evals November 2023 (1)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">cbde84b2-3937-4512-8d71-fee6c23e2ec6</guid>
      <link>https://kairos.fm/intoaisafety/e007</link>
      <description>
        <![CDATA[<p>This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of the day, turned out to be pretty good.</p><p><br></p><p>Check out <a href="https://apartresearch.com">Apart Research</a>, the group that runs the <a href="https://alignmentjam.com">AlignmentJamz Hackathons</a>.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://arxiv.org/abs/2311.07723">Generalization Analogies: A Testbed for Generalizing AI Oversight to Hard-To-Measure Domains</a><ul><li><a href="https://www.lesswrong.com/posts/Yio4nmD8JMttx9o9S/new-paper-shows-truthfulness-and-instruction-following-don-t">New paper shows truthfulness &amp; instruction-following don't generalize by default</a></li><li><a href="https://joshuaclymer.github.io/generalization-analogies-website/">Generalization Analogies Website</a></li></ul></li><li><a href="https://arxiv.org/abs/2212.09251">Discovering Language Model Behaviors with Model-Written Evaluations</a><ul><li><a href="https://www.evals.anthropic.com">Model-Written Evals Website</a></li></ul></li><li><a href="https://github.com/openai/evals/tree/main/evals/elsuite/make_me_say">OpenAI Evals GitHub</a></li><li><a href="https://metr.org%20">METR</a> (previously ARC Evals)</li><li><a href="https://en.wikipedia.org/wiki/Goodhart%27s_law">Goodharting on Wikipedia</a></li><li><a href="https://arxiv.org/abs/2308.12014">From Instructions to Intrinsic Human Values, a Survey of Alignment Goals for Big Models</a></li><li><a href="https://arxiv.org/abs/2310.03693">Fine Tuning Aligned Language Models Compromises Safety Even When Users Do Not Intend</a></li><li><a href="https://arxiv.org/abs/2310.02949">Shadow Alignment: The Ease of Subverting Safely Aligned Language Models</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will Releasing the Weights of Future Large Language Models Grant Widespread Access to Pandemic Agents?</a></li><li><a href="https://www.sciencedirect.com/science/article/pii/S2666389923002210">Building Less Flawed Metrics, Understanding and Creating Better Measurement and Incentive Systems</a></li><li><a href="https://github.com/EleutherAI/lm-evaluation-harness">eLeutherAI's Model Evaluation Harness</a></li><li><a href="https://github.com/danbraunai/evalugator/tree/main">Evalugator Library</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of the day, turned out to be pretty good.</p><p><br></p><p>Check out <a href="https://apartresearch.com">Apart Research</a>, the group that runs the <a href="https://alignmentjam.com">AlignmentJamz Hackathons</a>.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://arxiv.org/abs/2311.07723">Generalization Analogies: A Testbed for Generalizing AI Oversight to Hard-To-Measure Domains</a><ul><li><a href="https://www.lesswrong.com/posts/Yio4nmD8JMttx9o9S/new-paper-shows-truthfulness-and-instruction-following-don-t">New paper shows truthfulness &amp; instruction-following don't generalize by default</a></li><li><a href="https://joshuaclymer.github.io/generalization-analogies-website/">Generalization Analogies Website</a></li></ul></li><li><a href="https://arxiv.org/abs/2212.09251">Discovering Language Model Behaviors with Model-Written Evaluations</a><ul><li><a href="https://www.evals.anthropic.com">Model-Written Evals Website</a></li></ul></li><li><a href="https://github.com/openai/evals/tree/main/evals/elsuite/make_me_say">OpenAI Evals GitHub</a></li><li><a href="https://metr.org%20">METR</a> (previously ARC Evals)</li><li><a href="https://en.wikipedia.org/wiki/Goodhart%27s_law">Goodharting on Wikipedia</a></li><li><a href="https://arxiv.org/abs/2308.12014">From Instructions to Intrinsic Human Values, a Survey of Alignment Goals for Big Models</a></li><li><a href="https://arxiv.org/abs/2310.03693">Fine Tuning Aligned Language Models Compromises Safety Even When Users Do Not Intend</a></li><li><a href="https://arxiv.org/abs/2310.02949">Shadow Alignment: The Ease of Subverting Safely Aligned Language Models</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will Releasing the Weights of Future Large Language Models Grant Widespread Access to Pandemic Agents?</a></li><li><a href="https://www.sciencedirect.com/science/article/pii/S2666389923002210">Building Less Flawed Metrics, Understanding and Creating Better Measurement and Incentive Systems</a></li><li><a href="https://github.com/EleutherAI/lm-evaluation-harness">eLeutherAI's Model Evaluation Harness</a></li><li><a href="https://github.com/danbraunai/evalugator/tree/main">Evalugator Library</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 08 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/04e9b1a9/e76d06ac.mp3" length="65892090" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>4119</itunes:duration>
      <itunes:summary>This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of the day, turned out to be pretty good.Check out Apart Research, the group that runs the AlignmentJamz Hackathons.Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.Generalization Analogies: A Testbed for Generalizing AI Oversight to Hard-To-Measure DomainsNew paper shows truthfulness &amp;amp; instruction-following don't generalize by defaultGeneralization Analogies WebsiteDiscovering Language Model Behaviors with Model-Written EvaluationsModel-Written Evals WebsiteOpenAI Evals GitHubMETR (previously ARC Evals)Goodharting on WikipediaFrom Instructions to Intrinsic Human Values, a Survey of Alignment Goals for Big ModelsFine Tuning Aligned Language Models Compromises Safety Even When Users Do Not IntendShadow Alignment: The Ease of Subverting Safely Aligned Language ModelsWill Releasing the Weights of Future Large Language Models Grant Widespread Access to Pandemic Agents?Building Less Flawed Metrics, Understanding and Creating Better Measurement and Incentive SystemseLeutherAI's Model Evaluation HarnessEvalugator Library</itunes:summary>
      <itunes:subtitle>This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of th</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Staying Up-to-Date in AI</title>
      <itunes:episode>6</itunes:episode>
      <podcast:episode>6</podcast:episode>
      <itunes:title>MINISODE: Staying Up-to-Date in AI</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">affbff67-b344-480f-bdf4-d62c1ee3e82c</guid>
      <link>https://kairos.fm/intoaisafety/e006</link>
      <description>
        <![CDATA[<p>In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the future.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li>Tools<ul><li><a href="https://feedly.com">Feedly</a></li><li><a href="https://arxiv-sanity-lite.com">arXiv Sanity Lite</a></li><li><a href="https://www.zotero.org">Zotero</a></li><li><a href="https://alternativeto.net">AlternativeTo</a></li></ul></li><li>My "Distilled AI" Folder<ul><li><a href="https://www.youtube.com/channel/UCNJ1Ymd5yFuUPtn21xtRbbw">AI Explained</a> YouTube channel</li><li><a href="https://newsletter.safe.ai">AI Safety newsletter</a></li><li><a href="https://datamachina.com">Data Machina</a> newsletter</li><li><a href="https://importai.substack.com">Import AI</a></li><li><a href="https://midwitalignment.substack.com">Midwit Alignment</a></li></ul></li><li>Honourable Mentions<ul><li><a href="https://www.alignmentforum.org">AI Alignment Forum</a></li><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://bounded-regret.ghost.io">Bounded Regret</a> (Jacob Steinhart's blog)</li><li><a href="https://www.cold-takes.com">Cold Takes</a> (Holden Karnofsky's blog)</li><li><a href="https://colah.github.io">Chris Olah's blog</a></li><li><a href="https://timdettmers.com">Tim Dettmers blog</a></li><li><a href="https://epochai.org/blog">Epoch blog</a></li><li><a href="https://www.apolloresearch.ai/blog">Apollo Research blog</a></li></ul></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the future.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li>Tools<ul><li><a href="https://feedly.com">Feedly</a></li><li><a href="https://arxiv-sanity-lite.com">arXiv Sanity Lite</a></li><li><a href="https://www.zotero.org">Zotero</a></li><li><a href="https://alternativeto.net">AlternativeTo</a></li></ul></li><li>My "Distilled AI" Folder<ul><li><a href="https://www.youtube.com/channel/UCNJ1Ymd5yFuUPtn21xtRbbw">AI Explained</a> YouTube channel</li><li><a href="https://newsletter.safe.ai">AI Safety newsletter</a></li><li><a href="https://datamachina.com">Data Machina</a> newsletter</li><li><a href="https://importai.substack.com">Import AI</a></li><li><a href="https://midwitalignment.substack.com">Midwit Alignment</a></li></ul></li><li>Honourable Mentions<ul><li><a href="https://www.alignmentforum.org">AI Alignment Forum</a></li><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://bounded-regret.ghost.io">Bounded Regret</a> (Jacob Steinhart's blog)</li><li><a href="https://www.cold-takes.com">Cold Takes</a> (Holden Karnofsky's blog)</li><li><a href="https://colah.github.io">Chris Olah's blog</a></li><li><a href="https://timdettmers.com">Tim Dettmers blog</a></li><li><a href="https://epochai.org/blog">Epoch blog</a></li><li><a href="https://www.apolloresearch.ai/blog">Apollo Research blog</a></li></ul></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 01 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/8dc3abe2/7c94b204.mp3" length="12554492" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>785</itunes:duration>
      <itunes:summary>In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the future.

Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.
ToolsFeedlyarXiv Sanity LiteZoteroAlternativeToMy "Distilled AI" FolderAI Explained YouTube channelAI Safety newsletterData Machina newsletterImport AIMidwit AlignmentHonourable MentionsAI Alignment ForumLessWrongBounded Regret (Jacob Steinhart's blog)Cold Takes (Holden Karnofsky's blog)Chris Olah's blog Tim Dettmers blogEpoch blogApollo Research blog</itunes:summary>
      <itunes:subtitle>In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the futur</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>INTERVIEW: Applications w/ Alice Rigg</title>
      <itunes:episode>5</itunes:episode>
      <podcast:episode>5</podcast:episode>
      <itunes:title>INTERVIEW: Applications w/ Alice Rigg</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">b2215bea-18f8-497a-b76f-4bb095c79cf0</guid>
      <link>https://kairos.fm/intoaisafety/e005</link>
      <description>
        <![CDATA[<p>Alice Rigg, a mechanistic interpretability researcher from Ottawa, Canada, joins me to discuss their path and the applications process for research/mentorship programs.</p><p>Join the <a href="https://discord.gg/KNxJwaytTG">Mech Interp Discord server</a> and attend reading groups at 11:00am on Wednesdays (Mountain Time)!</p><p>Check out Alice's <a href="https://woog97.github.io">website</a>.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://www.eleuther.ai">EleutherAI</a><br> <ul><li>Join the public <a href="https://www.discord.gg/eleutherai">EleutherAI discord server</a></li></ul></li><li> </li><li><a href="https://distill.pub">Distill</a></li><li><a href="https://www.effectivealtruism.org">Effective Altruism</a> (EA)</li><li><a href="https://www.lesswrong.com/posts/zwf68YaySvXhWYCdh/mats-summer-2023-retrospective">MATS Retrospective Summer 2023</a> post</li><li><a href="https://docs.google.com/document/d/1jce3f64Fz7PXmdCEyd9i0PTmcFaiP1pZdcBn5ye5sxY/edit">Ambitious Mechanistic Interpretability</a> AISC research plan by Alice Rigg</li><li><a href="https://berkeleyaisafety.com/spar">SPAR</a></li><li><a href="https://stability.ai">Stability AI</a><br> <ul><li>During their most recent fundraising round, Stability AI had a valuation of $4B <a href="https://www.bloomberg.com/news/articles/2023-03-03/openai-rival-stable-diffusion-maker-seeks-to-raise-funds-at-4-billion-valuation">(Bloomberg)</a></li></ul></li><li> </li><li><a href="https://discord.gg/KNxJwaytTG">Mech Interp Discord Server</a></li></ul><p><br></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Alice Rigg, a mechanistic interpretability researcher from Ottawa, Canada, joins me to discuss their path and the applications process for research/mentorship programs.</p><p>Join the <a href="https://discord.gg/KNxJwaytTG">Mech Interp Discord server</a> and attend reading groups at 11:00am on Wednesdays (Mountain Time)!</p><p>Check out Alice's <a href="https://woog97.github.io">website</a>.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://www.eleuther.ai">EleutherAI</a><br> <ul><li>Join the public <a href="https://www.discord.gg/eleutherai">EleutherAI discord server</a></li></ul></li><li> </li><li><a href="https://distill.pub">Distill</a></li><li><a href="https://www.effectivealtruism.org">Effective Altruism</a> (EA)</li><li><a href="https://www.lesswrong.com/posts/zwf68YaySvXhWYCdh/mats-summer-2023-retrospective">MATS Retrospective Summer 2023</a> post</li><li><a href="https://docs.google.com/document/d/1jce3f64Fz7PXmdCEyd9i0PTmcFaiP1pZdcBn5ye5sxY/edit">Ambitious Mechanistic Interpretability</a> AISC research plan by Alice Rigg</li><li><a href="https://berkeleyaisafety.com/spar">SPAR</a></li><li><a href="https://stability.ai">Stability AI</a><br> <ul><li>During their most recent fundraising round, Stability AI had a valuation of $4B <a href="https://www.bloomberg.com/news/articles/2023-03-03/openai-rival-stable-diffusion-maker-seeks-to-raise-funds-at-4-billion-valuation">(Bloomberg)</a></li></ul></li><li> </li><li><a href="https://discord.gg/KNxJwaytTG">Mech Interp Discord Server</a></li></ul><p><br></p>]]>
      </content:encoded>
      <pubDate>Mon, 18 Dec 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/2889532d/02620130.mp3" length="67843965" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>4241</itunes:duration>
      <itunes:summary>Alice Rigg, a mechanistic interpretability researcher from Ottawa, Canada, joins me to discuss their path and the applications process for research/mentorship programs.
Join the Mech Interp Discord server and attend reading groups at 11:00am on Wednesdays (Mountain Time)!
Check out Alice's website.
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. 

 EleutherAI
 Join the public EleutherAI discord server

Distill
 Effective Altruism (EA)
 MATS Retrospective Summer 2023 post
 Ambitious Mechanistic Interpretability AISC research plan by Alice Rigg
 SPAR
  Stability AI
  During their most recent fundraising round, Stability AI had a valuation of $4B (Bloomberg)

Mech Interp Discord Server</itunes:summary>
      <itunes:subtitle>Alice Rigg, a mechanistic interpretability researcher from Ottawa, Canada, joins me to discuss their path and the applications process for research/mentorship programs.
Join the Mech Interp Discord server and attend reading groups at 11:00am on Wednesday</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Program Applications (Winter 2024)</title>
      <itunes:episode>4</itunes:episode>
      <podcast:episode>4</podcast:episode>
      <itunes:title>MINISODE: Program Applications (Winter 2024)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">2f5c7fd0-e665-4850-a48c-b4d20f05f4af</guid>
      <link>https://kairos.fm/intoaisafety/e004</link>
      <description>
        <![CDATA[<p>We're back after a month-long hiatus with a podcast refactor and advice on the applications process for research/mentorship programs.</p><p>Check out the <a href="https://into-ai-safety.github.io/about/">About</a> page on the <a href="https://into-ai-safety.github.io">Into AI Safety</a> website for a summary of the logistics updates.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://www.matsprogram.org">MATS</a></li><li><a href="https://www.constellation.org/programs/astra-fellowship">ASTRA Fellowship</a></li><li><a href="https://www.arena.education">ARENA</a></li><li><a href="https://aisafety.camp">AI Safety Camp</a></li><li><a href="https://www.bluedotimpact.org">BlueDot Impact</a></li><li><a href="https://www.techwithtim.net">Tech with Tim</a></li><li>Fast.AI's <a href="https://course.fast.ai">Practical Deep Learning for Coders</a></li><li><a href="https://www.kaggle.com">Kaggle</a></li><li><a href="https://alignmentjam.com">AlignmentJams</a></li><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://www.alignmentforum.org">AI Alignment Forum</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>We're back after a month-long hiatus with a podcast refactor and advice on the applications process for research/mentorship programs.</p><p>Check out the <a href="https://into-ai-safety.github.io/about/">About</a> page on the <a href="https://into-ai-safety.github.io">Into AI Safety</a> website for a summary of the logistics updates.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://www.matsprogram.org">MATS</a></li><li><a href="https://www.constellation.org/programs/astra-fellowship">ASTRA Fellowship</a></li><li><a href="https://www.arena.education">ARENA</a></li><li><a href="https://aisafety.camp">AI Safety Camp</a></li><li><a href="https://www.bluedotimpact.org">BlueDot Impact</a></li><li><a href="https://www.techwithtim.net">Tech with Tim</a></li><li>Fast.AI's <a href="https://course.fast.ai">Practical Deep Learning for Coders</a></li><li><a href="https://www.kaggle.com">Kaggle</a></li><li><a href="https://alignmentjam.com">AlignmentJams</a></li><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://www.alignmentforum.org">AI Alignment Forum</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 11 Dec 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/6a54c881/0ef512b8.mp3" length="17356182" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>1085</itunes:duration>
      <itunes:summary>We're back after a month-long hiatus with a podcast refactor and advice on the applications process for research/mentorship programs.
Check out the About page on the Into AI Safety website for a summary of the logistics updates.
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. MATSASTRA FellowshipARENAAI Safety Camp
 BlueDot ImpactTech with TimFast.AI's Practical Deep Learning for CodersKaggleAlignmentJamsLessWrongAI Alignment Forum</itunes:summary>
      <itunes:subtitle>We're back after a month-long hiatus with a podcast refactor and advice on the applications process for research/mentorship programs.
Check out the About page on the Into AI Safety website for a summary of the logistics updates.
Links to all articles/pa</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: EAG Takeaways (Boston 2023)</title>
      <itunes:episode>3</itunes:episode>
      <podcast:episode>3</podcast:episode>
      <itunes:title>MINISODE: EAG Takeaways (Boston 2023)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">dc2fc55a-dca7-44c6-bf9e-be0c4e9eb971</guid>
      <link>https://kairos.fm/intoaisafety/e003</link>
      <description>
        <![CDATA[<p>This episode is a brief overview of the major takeaways I had from attending EAG Boston 2023, and an update on my plans for the podcast moving forward.</p><p><strong>TL;DL</strong></p><ul><li>Starting in early December (2023), I will be uploading episodes on a biweekly basis (day TBD).</li><li>I won't be releasing another episode until then, so that I can build a cache of episodes up.</li><li>During this month (November 2023), I'll also try to get the podcast up on more platforms, set up comments on more platforms, and create an anonymous feedback form.</li></ul><p><strong>Links</strong> Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://forum.effectivealtruism.org/posts/8R2NffQiCsn3F7hpv/how-to-generate-research-proposals">How to generate research proposals</a></li><li><a href="https://www.effectivealtruism.org/articles/karolina-sarek-how-to-do-research-that-matters">Karolina Sarek: How to do research that matters</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will releasing the weights of future large language models grant widespread access to pandemic agents?</a></li></ul><p>Like the show? Think it could be improved? Fill out <a href="https://docs.google.com/forms/d/e/1FAIpQLSdXqpGGb0uWgpQd8CUvKP6g2Ki8FrDsEBzFfQBrBoSZPlxjDQ/viewform?usp=sf_link">this anonymous feedback form </a>to let me know!</p><p>Please email all inquiries to intoaisafety@gmail.com.</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>This episode is a brief overview of the major takeaways I had from attending EAG Boston 2023, and an update on my plans for the podcast moving forward.</p><p><strong>TL;DL</strong></p><ul><li>Starting in early December (2023), I will be uploading episodes on a biweekly basis (day TBD).</li><li>I won't be releasing another episode until then, so that I can build a cache of episodes up.</li><li>During this month (November 2023), I'll also try to get the podcast up on more platforms, set up comments on more platforms, and create an anonymous feedback form.</li></ul><p><strong>Links</strong> Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://forum.effectivealtruism.org/posts/8R2NffQiCsn3F7hpv/how-to-generate-research-proposals">How to generate research proposals</a></li><li><a href="https://www.effectivealtruism.org/articles/karolina-sarek-how-to-do-research-that-matters">Karolina Sarek: How to do research that matters</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will releasing the weights of future large language models grant widespread access to pandemic agents?</a></li></ul><p>Like the show? Think it could be improved? Fill out <a href="https://docs.google.com/forms/d/e/1FAIpQLSdXqpGGb0uWgpQd8CUvKP6g2Ki8FrDsEBzFfQBrBoSZPlxjDQ/viewform?usp=sf_link">this anonymous feedback form </a>to let me know!</p><p>Please email all inquiries to intoaisafety@gmail.com.</p>]]>
      </content:encoded>
      <pubDate>Mon, 04 Dec 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/101ea918/e54e49cc.mp3" length="9228951" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>577</itunes:duration>
      <itunes:summary>This episode is a brief overview of the major takeaways I had from attending EAG Boston 2023, and an update on my plans for the podcast moving forward.
TL;DL Starting in early December (2023), I will be uploading episodes on a biweekly basis (day TBD).I won't be releasing another episode until then, so that I can build a cache of episodes up.During this month (November 2023), I'll also try to get the podcast up on more platforms, set up comments on more platforms, and create an anonymous feedback form.
Links
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. How to generate research proposalsKarolina Sarek: How to do research that mattersWill releasing the weights of future large language models grant widespread access to pandemic agents?

Like the show? Think it could be improved? Fill out this anonymous feedback form to let me know!
Please email all inquiries to intoaisafety@gmail.com.</itunes:summary>
      <itunes:subtitle>This episode is a brief overview of the major takeaways I had from attending EAG Boston 2023, and an update on my plans for the podcast moving forward.
TL;DL Starting in early December (2023), I will be uploading episodes on a biweekly basis (day TBD).I </itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
    </item>
    <item>
      <title>FEEDBACK: AISC Proposal w/ Remmelt Ellen</title>
      <itunes:episode>2</itunes:episode>
      <podcast:episode>2</podcast:episode>
      <itunes:title>FEEDBACK: AISC Proposal w/ Remmelt Ellen</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">084356f7-066c-49d9-b8d9-a45f7162f138</guid>
      <link>https://kairos.fm/intoaisafety/e002</link>
      <description>
        <![CDATA[<p>In this episode I discuss my initial research proposal for the 2024 Winter AI Safety Camp with one of the individuals who helps facilitate the program, Remmelt Ellen.</p><p>The proposal is titled The Effect of Machine Learning on Bioengineered Pandemic Risk. A doc-capsule of the proposal at the time of this recording can be found <a href="https://docs.google.com/document/d/1bbFDNc_hzhzYqN6pn1jYvPhSRp2CyYez4UP0ly_mMRM/edit?usp=sharing">at this link</a>.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://chemrxiv.org/engage/chemrxiv/article-details/61551803d1fc335b7cf8fd45">MegaSyn: Integrating Generative Molecule Design, Automated Analog Designer and Synthetic Viability Prediction</a></li><li><a href="https://www.nature.com/articles/s42256-022-00465-9?fbclid=IwAR11_V1cd9SUxEvUfwrWMA7TUcroyYIY1nBDUL3KaS-8B4rG5MIqZCmjm0M">Dual use of artificial-intelligence-powered drug discovery</a></li><li><a href="https://arxiv.org/abs/2306.13952">Artificial intelligence and biological misuse: Differentiating risks of language models and biological design tools</a></li><li><a href="https://www.alignmentforum.org/posts/ChDH335ckdvpxXaXX/model-organisms-of-misalignment-the-case-for-a-new-pillar-of-1">Model Organisms of Misalignment: The Case for a New Pillar of Alignment Research</a></li><li><a href="https://arxiv.org/abs/2310.02949">Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models</a></li><li><a href="https://arxiv.org/abs/2310.03693">Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!</a></li><li><a href="https://www.alignmentforum.org/posts/3eqHYxfWb5x4Qfz8C/unrlhf-efficiently-undoing-llm-safeguards">unRLHF - Efficiently undoing LLM safeguards</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode I discuss my initial research proposal for the 2024 Winter AI Safety Camp with one of the individuals who helps facilitate the program, Remmelt Ellen.</p><p>The proposal is titled The Effect of Machine Learning on Bioengineered Pandemic Risk. A doc-capsule of the proposal at the time of this recording can be found <a href="https://docs.google.com/document/d/1bbFDNc_hzhzYqN6pn1jYvPhSRp2CyYez4UP0ly_mMRM/edit?usp=sharing">at this link</a>.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://chemrxiv.org/engage/chemrxiv/article-details/61551803d1fc335b7cf8fd45">MegaSyn: Integrating Generative Molecule Design, Automated Analog Designer and Synthetic Viability Prediction</a></li><li><a href="https://www.nature.com/articles/s42256-022-00465-9?fbclid=IwAR11_V1cd9SUxEvUfwrWMA7TUcroyYIY1nBDUL3KaS-8B4rG5MIqZCmjm0M">Dual use of artificial-intelligence-powered drug discovery</a></li><li><a href="https://arxiv.org/abs/2306.13952">Artificial intelligence and biological misuse: Differentiating risks of language models and biological design tools</a></li><li><a href="https://www.alignmentforum.org/posts/ChDH335ckdvpxXaXX/model-organisms-of-misalignment-the-case-for-a-new-pillar-of-1">Model Organisms of Misalignment: The Case for a New Pillar of Alignment Research</a></li><li><a href="https://arxiv.org/abs/2310.02949">Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models</a></li><li><a href="https://arxiv.org/abs/2310.03693">Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!</a></li><li><a href="https://www.alignmentforum.org/posts/3eqHYxfWb5x4Qfz8C/unrlhf-efficiently-undoing-llm-safeguards">unRLHF - Efficiently undoing LLM safeguards</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 27 Nov 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/57154a54/62466372.mp3" length="54368967" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>3398</itunes:duration>
      <itunes:summary>In this episode I discuss my initial research proposal for the 2024 Winter AI Safety Camp with one of the individuals who helps facilitate the program, Remmelt Ellen.
The proposal is titled The Effect of Machine Learning on Bioengineered Pandemic Risk. A doc-capsule of the proposal at the time of this recording can be found at this link.
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. MegaSyn: Integrating Generative Molecule Design, Automated Analog Designer and Synthetic Viability PredictionDual use of artificial-intelligence-powered drug discoveryArtificial intelligence and biological misuse: Differentiating risks of language models and biological design toolsModel Organisms of Misalignment: The Case for a New Pillar of Alignment ResearchShadow Alignment: The Ease of Subverting Safely-Aligned Language ModelsFine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!unRLHF - Efficiently undoing LLM safeguards</itunes:summary>
      <itunes:subtitle>In this episode I discuss my initial research proposal for the 2024 Winter AI Safety Camp with one of the individuals who helps facilitate the program, Remmelt Ellen.
The proposal is titled The Effect of Machine Learning on Bioengineered Pandemic Risk. A</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
    </item>
    <item>
      <title>MINISODE: Introduction and Motivation</title>
      <itunes:episode>1</itunes:episode>
      <podcast:episode>1</podcast:episode>
      <itunes:title>MINISODE: Introduction and Motivation</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">4a709b26-c999-4113-b23f-daeac781b2df</guid>
      <link>https://kairos.fm/intoaisafety/e001</link>
      <description>
        <![CDATA[<p>Welcome to the <strong>Into AI Safety</strong> podcast! In this episode I provide reasoning for why I am starting this podcast, what I am trying to accomplish with it, and a little bit of background on how I got here.</p><p>Please email all inquiries and suggestions to intoaisafety@gmail.com.</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Welcome to the <strong>Into AI Safety</strong> podcast! In this episode I provide reasoning for why I am starting this podcast, what I am trying to accomplish with it, and a little bit of background on how I got here.</p><p>Please email all inquiries and suggestions to intoaisafety@gmail.com.</p>]]>
      </content:encoded>
      <pubDate>Mon, 13 Nov 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/2b9dbb68/7bb95cf1.mp3" length="9326333" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>583</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Welcome to the <strong>Into AI Safety</strong> podcast! In this episode I provide reasoning for why I am starting this podcast, what I am trying to accomplish with it, and a little bit of background on how I got here.</p><p>Please email all inquiries and suggestions to intoaisafety@gmail.com.</p>]]>
      </itunes:summary>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistorcdn.com/wCpjhOlh6OKmPqoVVkuQKXFSzSWrIsUwtdSqaT2zb2M/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
    </item>
  </channel>
</rss>
