<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet href="https://feeds.transistor.fm/stylesheet.xsl" type="text/xsl"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:podcast="https://podcastindex.org/namespace/1.0">
  <channel>
    <atom:link rel="self" type="application/atom+xml" href="https://feeds.transistor.fm/intoaisafety" title="MP3 Audio"/>
    <atom:link rel="hub" href="https://pubsubhubbub.appspot.com/"/>
    <podcast:podping usesPodping="true"/>
    <title>Into AI Safety</title>
    <generator>Transistor (https://transistor.fm)</generator>
    <itunes:new-feed-url>https://feeds.transistor.fm/intoaisafety</itunes:new-feed-url>
    <description>The Into AI Safety podcast aims to make it easier for everyone, regardless of background, to get meaningfully involved with the conversations surrounding the rules and regulations which should govern the research, development, deployment, and use of the technologies encompassed by the term "artificial intelligence" or "AI"

For better formatted show notes, additional resources, and more, go to https://into-ai-safety.github.io
For even more content and community engagement, head over to my Patreon at https://www.patreon.com/IntoAISafety</description>
    <copyright>© Kairos.fm</copyright>
    <podcast:locked owner="listen@kairos.fm">no</podcast:locked>
    <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
    <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    <language>en-us</language>
    <pubDate>Mon, 30 Dec 2024 17:19:55 -0700</pubDate>
    <lastBuildDate>Mon, 30 Dec 2024 17:20:09 -0700</lastBuildDate>
    <link>https://kairos.fm/intoaisafety/</link>
    <image>
      <url>https://img.transistor.fm/Dx6WjxGUnscbtCfCV6aiRIOyUf4pu56y9syxIutOKcE/rs:fill:3000:3000:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8yOTE5/MTU3M2U5NDAzYjJk/OTcwZTc5NzNkYzE3/ZWFjNC5qcGVn.jpg</url>
      <title>Into AI Safety</title>
      <link>https://kairos.fm/intoaisafety/</link>
    </image>
    <itunes:category text="Technology"/>
    <itunes:category text="Science">
      <itunes:category text="Mathematics"/>
    </itunes:category>
    <itunes:type>episodic</itunes:type>
    <itunes:author>Jacob Haimes</itunes:author>
    <itunes:image href="https://img.transistor.fm/Dx6WjxGUnscbtCfCV6aiRIOyUf4pu56y9syxIutOKcE/rs:fill:3000:3000:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8yOTE5/MTU3M2U5NDAzYjJk/OTcwZTc5NzNkYzE3/ZWFjNC5qcGVn.jpg"/>
    <itunes:summary>The Into AI Safety podcast aims to make it easier for everyone, regardless of background, to get meaningfully involved with the conversations surrounding the rules and regulations which should govern the research, development, deployment, and use of the technologies encompassed by the term "artificial intelligence" or "AI"

For better formatted show notes, additional resources, and more, go to https://into-ai-safety.github.io
For even more content and community engagement, head over to my Patreon at https://www.patreon.com/IntoAISafety</itunes:summary>
    <itunes:subtitle>The Into AI Safety podcast aims to make it easier for everyone, regardless of background, to get meaningfully involved with the conversations surrounding the rules and regulations which should govern the research, development, deployment, and use of the technologies encompassed by the term "artificial intelligence" or "AI"

For better formatted show notes, additional resources, and more, go to https://into-ai-safety.github.io
For even more content and community engagement, head over to my Patreon at https://www.patreon.com/IntoAISafety.</itunes:subtitle>
    <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
    <itunes:owner>
      <itunes:name>Jacob Haimes</itunes:name>
      <itunes:email>listen@kairos.fm</itunes:email>
    </itunes:owner>
    <itunes:complete>No</itunes:complete>
    <itunes:explicit>No</itunes:explicit>
    <item>
      <title>INTERVIEW: Scaling Democracy w/ (Dr.) Igor Krawczuk</title>
      <itunes:episode>19</itunes:episode>
      <podcast:episode>19</podcast:episode>
      <itunes:title>INTERVIEW: Scaling Democracy w/ (Dr.) Igor Krawczuk</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">81c52538-5af2-40aa-abd7-02c08088f43e</guid>
      <link>https://kairos.fm/intoaisafety/e019</link>
      <description>
        <![CDATA[<p>The <em>almost</em> Dr. Igor Krawczuk joins me for what is the equivalent of 4 of my previous episodes. We get into all the classics: eugenics, capitalism, philosophical toads... Need I say more?</p><p>If you're interested in connecting with Igor, head on over to his <a href="https://krawczuk.eu/">website</a>, or check out <a href="https://github.com/into-ai-safety/into-ai-safety.github.io/blob/master/_posts">placeholder for thesis</a> (it isn't published yet).</p><p>Because the full show notes have a whopping 115 additional links, I'll highlight some that I think are particularly worthwhile here:</p><ul><li>The best article you'll ever read on <a href="https://jacob-haimes.github.io/independent/Open-Source-AI-is-a-lie/">Open Source AI</a></li><li>The best article you'll ever read on <a href="https://www.odysseaninstitute.org/post/let-s-talk-about-emergence">emergence in ML</a></li><li>Kate Crawford's <a href="https://yalebooks.yale.edu/book/9780300264630/atlas-of-ai/"><em>Atlas of AI</em></a> (<a href="https://en.wikipedia.org/wiki/Atlas_of_AI">Wikipedia</a>)</li><li><a href="https://arxiv.org/abs/1911.01547">On the Measure of Intelligence</a></li><li>Thomas Piketty's <a href="https://www.hup.harvard.edu/books/9780674430006"><em>Capital in the Twenty-First Century</em></a> (<a href="https://en.wikipedia.org/wiki/Capital_in_the_Twenty-First_Century">Wikipedia</a>)</li><li>Yurii Nesterov's <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=2-ElBQAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;dq=info:JTiRBrZ_LZMJ:scholar.google.com&amp;ots=wnpRdsxjjv&amp;sig=1Oa-5P-zZZ_MX_2MFKv5cq2fx48#v=onepage&amp;q&amp;f=false"><em>Introductory Lectures on Convex Optimization</em></a></li></ul><p><strong>Chapters<br></strong></p><ul><li>(02:32) - Introducing Igor</li>
<li>(10:11) - Aside on EY, LW, EA, etc., a.k.a. lettersoup</li>
<li>(18:30) - Igor on AI alignment</li>
<li>(33:06) - "Open Source" in AI</li>
<li>(41:20) - The story of infinite riches and suffering</li>
<li>(59:11) - On AI threat models</li>
<li>(01:09:25) - Representation in AI</li>
<li>(01:15:00) - Hazard fishing</li>
<li>(01:18:52) - Intelligence and eugenics</li>
<li>(01:34:38) - Emergence</li>
<li>(01:48:19) - Considering externalities</li>
<li>(01:53:33) - The shape of an argument</li>
<li>(02:01:39) - More eugenics</li>
<li>(02:06:09) - I'm convinced, what now?</li>
<li>(02:18:03) - AIxBio (round ??)</li>
<li>(02:29:09) - On open release of models</li>
<li>(02:40:28) - Data and copyright</li>
<li>(02:44:09) - Scientific accessibility and bullshit</li>
<li>(02:53:04) - Igor's point of view</li>
<li>(02:57:20) - Outro</li>
</ul><p><strong><br>Links</strong></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. All references, including those only mentioned in the extended version of this episode, are included.</p><ul><li><a href="https://www.lighthousereports.com/suspicion-machines-methodology/">Suspicious Machines Methodology</a>, referred to as the "Rotterdam Lighthouse Report" in the episode</li><li><a href="https://www.epfl.ch/labs/lions/">LIONS Lab</a> at EPFL</li><li>The <a href="https://pbs.twimg.com/media/D53Q_MYW4AA-wRK.jpg">meme</a> that Igor references</li><li><a href="https://arxiv.org/abs/2401.01869">On the Hardness of Learning Under Symmetries</a></li><li><a href="https://uvagedl.github.io/">Course</a> on the concept of equivariant deep learning</li><li>Aside on EY/EA/etc.<ul><li>Sources on Eliezer Yudkowski<ul><li><a href="https://encyclopedia.pub/entry/33978">Scholarly Community Encyclopedia</a></li><li><a href="https://time.com/collection/time100-ai/6309037/eliezer-yudkowsky/">TIME100 AI</a></li><li>Yudkowski's personal <a href="https://www.yudkowsky.net/">website</a></li><li><a href="https://en.wikipedia.org/wiki/Eliezer_Yudkowsky">EY Wikipedia</a></li><li><a href="https://whatshouldiread.fandom.com/wiki/Eliezer_Yudkowsky#cite_note-1">A Very Literary Wiki</a> -TIME article: <a href="https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/">Pausing AI Developments Isn’t Enough. We Need to Shut it All Down</a> documenting EY's ruminations of bombing datacenters; this comes up later in the episode but is included here because it about EY.</li></ul></li><li><a href="https://www.lesswrong.com/">LessWrong</a><ul><li><a href="https://en.wikipedia.org/wiki/LessWrong">LW Wikipedia</a></li></ul></li><li><a href="https://intelligence.org/">MIRI</a></li><li>Coverage on Nick Bostrom (being a racist)<ul><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/28/nick-bostrom-controversial-future-of-humanity-institute-closure-longtermism-affective-altruism">‘Eugenics on steroids’: the toxic and contested legacy of Oxford’s Future of Humanity Institute</a></li><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/19/oxford-future-of-humanity-institute-closes">Oxford shuts down institute run by Elon Musk-backed philosopher</a></li></ul></li><li>Investigative <a href="https://markfuentes1.substack.com/p/emile-p-torress-history-of-dishonesty">piece</a> on Émile Torres</li><li><a href="https://dl.acm.org/doi/10.1145/3442188.3445922">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜</a></li><li>NY Times article: <a href="https://www.nytimes.com/2019/11/11/technology/artificial-intelligence-bias.html">We Teach A.I. Systems Everything, Including Our Biases</a></li><li>NY Times article: <a href="https://www.nytimes.com/2020/12/03/technology/google-researcher-timnit-gebru.html">Google Researcher Says She Was Fired Over Paper Highlighting Bias in A.I.</a></li><li>Timnit Gebru's <a href="https://en.wikipedia.org/wiki/Timnit_Gebru">Wikipedia</a></li><li><a href="https://firstmonday.org/ojs/index.php/fm/article/view/13636">The TESCREAL Bundle: Eugenics and the Promise of Utopia through Artificial General Intelligence</a></li><li>Sources on the environmental impact of LLMs<ul><li><a href="https://analyticsindiamag.com/the-environmental-impact-of-llms/">The Environmental Impact of LLMs</a></li><li><a href="https://tinyml.substack.com/p/the-cost-of-inference-running-the">The Cost of Inference: Running the Models</a></li><li><a href="https://arxiv.org/abs/1906.02243">Energy and Policy Considerations for Deep Learning in NLP</a></li><li><a href="https://weareyard.com/insights/the-carbon-impact-of-ai-vs-search-engines">The Carbon Impact of AI vs Search Engines</a></li></ul></li></ul></li><li><a href="https://www.science.org/doi/full/10.1126/science.abi7176?casa_token=2txe0r_jjhQAAAAA%3ALJa__HZL9COyj9EUpdILZdtnMKLyggfFe7Zpvv0tNze62rLO0CoQHCCJiXfruxUeBLj3YBZ33F8OOv0u">Filling Gaps in Trustworthy Development of AI </a>(Igor is an author on this one)</li><li><a href="https://www.hindawi.com/journals/complexity/2022/8210732/">A Computational Turn in Policy Process Studies: Coevolving Network Dynamics of Policy Change</a></li><li><a href="https://proceedings.neurips.cc/paper_files/paper/2020/file/7e05d6f828574fbc975a896b25bb011e-Paper.pdf">The Smoothed Possibility of Social Choice</a>, an intro in social choice theory and how it overlaps with ML</li><li>Relating to Dan Hendrycks<ul><li><a href="https://arxiv.org/abs/2303.16200">Natural Selection Favors AIs over Humans</a><ul><li>"One easy-to-digest source to highlight what he gets wrong [is] <a href="https://pressbooks.calstate.edu/explorationsbioanth2/chapter/17/">Social and Biopolitical Dimensions of Evolutionary Thinking</a>" -Igor</li></ul></li><li><a href="https://www.aisafetybook.com/">Introduction to AI Safety, Ethics, and Society</a>, recently published textbook</li><li>"<a href="https://arxiv.org/pdf/2306.12001#page=10.19">Source</a> to the section [of this paper] that makes Dan one of my favs from that crowd." -Igor</li><li><a href="https://twitter.com/DanHendrycks/status/1710312043503321141">Twitter post</a> referenced in the episode&lt;...</li></ul></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>The <em>almost</em> Dr. Igor Krawczuk joins me for what is the equivalent of 4 of my previous episodes. We get into all the classics: eugenics, capitalism, philosophical toads... Need I say more?</p><p>If you're interested in connecting with Igor, head on over to his <a href="https://krawczuk.eu/">website</a>, or check out <a href="https://github.com/into-ai-safety/into-ai-safety.github.io/blob/master/_posts">placeholder for thesis</a> (it isn't published yet).</p><p>Because the full show notes have a whopping 115 additional links, I'll highlight some that I think are particularly worthwhile here:</p><ul><li>The best article you'll ever read on <a href="https://jacob-haimes.github.io/independent/Open-Source-AI-is-a-lie/">Open Source AI</a></li><li>The best article you'll ever read on <a href="https://www.odysseaninstitute.org/post/let-s-talk-about-emergence">emergence in ML</a></li><li>Kate Crawford's <a href="https://yalebooks.yale.edu/book/9780300264630/atlas-of-ai/"><em>Atlas of AI</em></a> (<a href="https://en.wikipedia.org/wiki/Atlas_of_AI">Wikipedia</a>)</li><li><a href="https://arxiv.org/abs/1911.01547">On the Measure of Intelligence</a></li><li>Thomas Piketty's <a href="https://www.hup.harvard.edu/books/9780674430006"><em>Capital in the Twenty-First Century</em></a> (<a href="https://en.wikipedia.org/wiki/Capital_in_the_Twenty-First_Century">Wikipedia</a>)</li><li>Yurii Nesterov's <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=2-ElBQAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;dq=info:JTiRBrZ_LZMJ:scholar.google.com&amp;ots=wnpRdsxjjv&amp;sig=1Oa-5P-zZZ_MX_2MFKv5cq2fx48#v=onepage&amp;q&amp;f=false"><em>Introductory Lectures on Convex Optimization</em></a></li></ul><p><strong>Chapters<br></strong></p><ul><li>(02:32) - Introducing Igor</li>
<li>(10:11) - Aside on EY, LW, EA, etc., a.k.a. lettersoup</li>
<li>(18:30) - Igor on AI alignment</li>
<li>(33:06) - "Open Source" in AI</li>
<li>(41:20) - The story of infinite riches and suffering</li>
<li>(59:11) - On AI threat models</li>
<li>(01:09:25) - Representation in AI</li>
<li>(01:15:00) - Hazard fishing</li>
<li>(01:18:52) - Intelligence and eugenics</li>
<li>(01:34:38) - Emergence</li>
<li>(01:48:19) - Considering externalities</li>
<li>(01:53:33) - The shape of an argument</li>
<li>(02:01:39) - More eugenics</li>
<li>(02:06:09) - I'm convinced, what now?</li>
<li>(02:18:03) - AIxBio (round ??)</li>
<li>(02:29:09) - On open release of models</li>
<li>(02:40:28) - Data and copyright</li>
<li>(02:44:09) - Scientific accessibility and bullshit</li>
<li>(02:53:04) - Igor's point of view</li>
<li>(02:57:20) - Outro</li>
</ul><p><strong><br>Links</strong></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. All references, including those only mentioned in the extended version of this episode, are included.</p><ul><li><a href="https://www.lighthousereports.com/suspicion-machines-methodology/">Suspicious Machines Methodology</a>, referred to as the "Rotterdam Lighthouse Report" in the episode</li><li><a href="https://www.epfl.ch/labs/lions/">LIONS Lab</a> at EPFL</li><li>The <a href="https://pbs.twimg.com/media/D53Q_MYW4AA-wRK.jpg">meme</a> that Igor references</li><li><a href="https://arxiv.org/abs/2401.01869">On the Hardness of Learning Under Symmetries</a></li><li><a href="https://uvagedl.github.io/">Course</a> on the concept of equivariant deep learning</li><li>Aside on EY/EA/etc.<ul><li>Sources on Eliezer Yudkowski<ul><li><a href="https://encyclopedia.pub/entry/33978">Scholarly Community Encyclopedia</a></li><li><a href="https://time.com/collection/time100-ai/6309037/eliezer-yudkowsky/">TIME100 AI</a></li><li>Yudkowski's personal <a href="https://www.yudkowsky.net/">website</a></li><li><a href="https://en.wikipedia.org/wiki/Eliezer_Yudkowsky">EY Wikipedia</a></li><li><a href="https://whatshouldiread.fandom.com/wiki/Eliezer_Yudkowsky#cite_note-1">A Very Literary Wiki</a> -TIME article: <a href="https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/">Pausing AI Developments Isn’t Enough. We Need to Shut it All Down</a> documenting EY's ruminations of bombing datacenters; this comes up later in the episode but is included here because it about EY.</li></ul></li><li><a href="https://www.lesswrong.com/">LessWrong</a><ul><li><a href="https://en.wikipedia.org/wiki/LessWrong">LW Wikipedia</a></li></ul></li><li><a href="https://intelligence.org/">MIRI</a></li><li>Coverage on Nick Bostrom (being a racist)<ul><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/28/nick-bostrom-controversial-future-of-humanity-institute-closure-longtermism-affective-altruism">‘Eugenics on steroids’: the toxic and contested legacy of Oxford’s Future of Humanity Institute</a></li><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/19/oxford-future-of-humanity-institute-closes">Oxford shuts down institute run by Elon Musk-backed philosopher</a></li></ul></li><li>Investigative <a href="https://markfuentes1.substack.com/p/emile-p-torress-history-of-dishonesty">piece</a> on Émile Torres</li><li><a href="https://dl.acm.org/doi/10.1145/3442188.3445922">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜</a></li><li>NY Times article: <a href="https://www.nytimes.com/2019/11/11/technology/artificial-intelligence-bias.html">We Teach A.I. Systems Everything, Including Our Biases</a></li><li>NY Times article: <a href="https://www.nytimes.com/2020/12/03/technology/google-researcher-timnit-gebru.html">Google Researcher Says She Was Fired Over Paper Highlighting Bias in A.I.</a></li><li>Timnit Gebru's <a href="https://en.wikipedia.org/wiki/Timnit_Gebru">Wikipedia</a></li><li><a href="https://firstmonday.org/ojs/index.php/fm/article/view/13636">The TESCREAL Bundle: Eugenics and the Promise of Utopia through Artificial General Intelligence</a></li><li>Sources on the environmental impact of LLMs<ul><li><a href="https://analyticsindiamag.com/the-environmental-impact-of-llms/">The Environmental Impact of LLMs</a></li><li><a href="https://tinyml.substack.com/p/the-cost-of-inference-running-the">The Cost of Inference: Running the Models</a></li><li><a href="https://arxiv.org/abs/1906.02243">Energy and Policy Considerations for Deep Learning in NLP</a></li><li><a href="https://weareyard.com/insights/the-carbon-impact-of-ai-vs-search-engines">The Carbon Impact of AI vs Search Engines</a></li></ul></li></ul></li><li><a href="https://www.science.org/doi/full/10.1126/science.abi7176?casa_token=2txe0r_jjhQAAAAA%3ALJa__HZL9COyj9EUpdILZdtnMKLyggfFe7Zpvv0tNze62rLO0CoQHCCJiXfruxUeBLj3YBZ33F8OOv0u">Filling Gaps in Trustworthy Development of AI </a>(Igor is an author on this one)</li><li><a href="https://www.hindawi.com/journals/complexity/2022/8210732/">A Computational Turn in Policy Process Studies: Coevolving Network Dynamics of Policy Change</a></li><li><a href="https://proceedings.neurips.cc/paper_files/paper/2020/file/7e05d6f828574fbc975a896b25bb011e-Paper.pdf">The Smoothed Possibility of Social Choice</a>, an intro in social choice theory and how it overlaps with ML</li><li>Relating to Dan Hendrycks<ul><li><a href="https://arxiv.org/abs/2303.16200">Natural Selection Favors AIs over Humans</a><ul><li>"One easy-to-digest source to highlight what he gets wrong [is] <a href="https://pressbooks.calstate.edu/explorationsbioanth2/chapter/17/">Social and Biopolitical Dimensions of Evolutionary Thinking</a>" -Igor</li></ul></li><li><a href="https://www.aisafetybook.com/">Introduction to AI Safety, Ethics, and Society</a>, recently published textbook</li><li>"<a href="https://arxiv.org/pdf/2306.12001#page=10.19">Source</a> to the section [of this paper] that makes Dan one of my favs from that crowd." -Igor</li><li><a href="https://twitter.com/DanHendrycks/status/1710312043503321141">Twitter post</a> referenced in the episode&lt;...</li></ul></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 03 Jun 2024 10:11:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/b8225038/75bd8800.mp3" length="171643841" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>10726</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>The <em>almost</em> Dr. Igor Krawczuk joins me for what is the equivalent of 4 of my previous episodes. We get into all the classics: eugenics, capitalism, philosophical toads... Need I say more?</p><p>If you're interested in connecting with Igor, head on over to his <a href="https://krawczuk.eu/">website</a>, or check out <a href="https://github.com/into-ai-safety/into-ai-safety.github.io/blob/master/_posts">placeholder for thesis</a> (it isn't published yet).</p><p>Because the full show notes have a whopping 115 additional links, I'll highlight some that I think are particularly worthwhile here:</p><ul><li>The best article you'll ever read on <a href="https://jacob-haimes.github.io/independent/Open-Source-AI-is-a-lie/">Open Source AI</a></li><li>The best article you'll ever read on <a href="https://www.odysseaninstitute.org/post/let-s-talk-about-emergence">emergence in ML</a></li><li>Kate Crawford's <a href="https://yalebooks.yale.edu/book/9780300264630/atlas-of-ai/"><em>Atlas of AI</em></a> (<a href="https://en.wikipedia.org/wiki/Atlas_of_AI">Wikipedia</a>)</li><li><a href="https://arxiv.org/abs/1911.01547">On the Measure of Intelligence</a></li><li>Thomas Piketty's <a href="https://www.hup.harvard.edu/books/9780674430006"><em>Capital in the Twenty-First Century</em></a> (<a href="https://en.wikipedia.org/wiki/Capital_in_the_Twenty-First_Century">Wikipedia</a>)</li><li>Yurii Nesterov's <a href="https://books.google.com/books?hl=en&amp;lr=&amp;id=2-ElBQAAQBAJ&amp;oi=fnd&amp;pg=PA1&amp;dq=info:JTiRBrZ_LZMJ:scholar.google.com&amp;ots=wnpRdsxjjv&amp;sig=1Oa-5P-zZZ_MX_2MFKv5cq2fx48#v=onepage&amp;q&amp;f=false"><em>Introductory Lectures on Convex Optimization</em></a></li></ul><p><strong>Chapters<br></strong></p><ul><li>(02:32) - Introducing Igor</li>
<li>(10:11) - Aside on EY, LW, EA, etc., a.k.a. lettersoup</li>
<li>(18:30) - Igor on AI alignment</li>
<li>(33:06) - "Open Source" in AI</li>
<li>(41:20) - The story of infinite riches and suffering</li>
<li>(59:11) - On AI threat models</li>
<li>(01:09:25) - Representation in AI</li>
<li>(01:15:00) - Hazard fishing</li>
<li>(01:18:52) - Intelligence and eugenics</li>
<li>(01:34:38) - Emergence</li>
<li>(01:48:19) - Considering externalities</li>
<li>(01:53:33) - The shape of an argument</li>
<li>(02:01:39) - More eugenics</li>
<li>(02:06:09) - I'm convinced, what now?</li>
<li>(02:18:03) - AIxBio (round ??)</li>
<li>(02:29:09) - On open release of models</li>
<li>(02:40:28) - Data and copyright</li>
<li>(02:44:09) - Scientific accessibility and bullshit</li>
<li>(02:53:04) - Igor's point of view</li>
<li>(02:57:20) - Outro</li>
</ul><p><strong><br>Links</strong></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. All references, including those only mentioned in the extended version of this episode, are included.</p><ul><li><a href="https://www.lighthousereports.com/suspicion-machines-methodology/">Suspicious Machines Methodology</a>, referred to as the "Rotterdam Lighthouse Report" in the episode</li><li><a href="https://www.epfl.ch/labs/lions/">LIONS Lab</a> at EPFL</li><li>The <a href="https://pbs.twimg.com/media/D53Q_MYW4AA-wRK.jpg">meme</a> that Igor references</li><li><a href="https://arxiv.org/abs/2401.01869">On the Hardness of Learning Under Symmetries</a></li><li><a href="https://uvagedl.github.io/">Course</a> on the concept of equivariant deep learning</li><li>Aside on EY/EA/etc.<ul><li>Sources on Eliezer Yudkowski<ul><li><a href="https://encyclopedia.pub/entry/33978">Scholarly Community Encyclopedia</a></li><li><a href="https://time.com/collection/time100-ai/6309037/eliezer-yudkowsky/">TIME100 AI</a></li><li>Yudkowski's personal <a href="https://www.yudkowsky.net/">website</a></li><li><a href="https://en.wikipedia.org/wiki/Eliezer_Yudkowsky">EY Wikipedia</a></li><li><a href="https://whatshouldiread.fandom.com/wiki/Eliezer_Yudkowsky#cite_note-1">A Very Literary Wiki</a> -TIME article: <a href="https://time.com/6266923/ai-eliezer-yudkowsky-open-letter-not-enough/">Pausing AI Developments Isn’t Enough. We Need to Shut it All Down</a> documenting EY's ruminations of bombing datacenters; this comes up later in the episode but is included here because it about EY.</li></ul></li><li><a href="https://www.lesswrong.com/">LessWrong</a><ul><li><a href="https://en.wikipedia.org/wiki/LessWrong">LW Wikipedia</a></li></ul></li><li><a href="https://intelligence.org/">MIRI</a></li><li>Coverage on Nick Bostrom (being a racist)<ul><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/28/nick-bostrom-controversial-future-of-humanity-institute-closure-longtermism-affective-altruism">‘Eugenics on steroids’: the toxic and contested legacy of Oxford’s Future of Humanity Institute</a></li><li>The Guardian article: <a href="https://www.theguardian.com/technology/2024/apr/19/oxford-future-of-humanity-institute-closes">Oxford shuts down institute run by Elon Musk-backed philosopher</a></li></ul></li><li>Investigative <a href="https://markfuentes1.substack.com/p/emile-p-torress-history-of-dishonesty">piece</a> on Émile Torres</li><li><a href="https://dl.acm.org/doi/10.1145/3442188.3445922">On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜</a></li><li>NY Times article: <a href="https://www.nytimes.com/2019/11/11/technology/artificial-intelligence-bias.html">We Teach A.I. Systems Everything, Including Our Biases</a></li><li>NY Times article: <a href="https://www.nytimes.com/2020/12/03/technology/google-researcher-timnit-gebru.html">Google Researcher Says She Was Fired Over Paper Highlighting Bias in A.I.</a></li><li>Timnit Gebru's <a href="https://en.wikipedia.org/wiki/Timnit_Gebru">Wikipedia</a></li><li><a href="https://firstmonday.org/ojs/index.php/fm/article/view/13636">The TESCREAL Bundle: Eugenics and the Promise of Utopia through Artificial General Intelligence</a></li><li>Sources on the environmental impact of LLMs<ul><li><a href="https://analyticsindiamag.com/the-environmental-impact-of-llms/">The Environmental Impact of LLMs</a></li><li><a href="https://tinyml.substack.com/p/the-cost-of-inference-running-the">The Cost of Inference: Running the Models</a></li><li><a href="https://arxiv.org/abs/1906.02243">Energy and Policy Considerations for Deep Learning in NLP</a></li><li><a href="https://weareyard.com/insights/the-carbon-impact-of-ai-vs-search-engines">The Carbon Impact of AI vs Search Engines</a></li></ul></li></ul></li><li><a href="https://www.science.org/doi/full/10.1126/science.abi7176?casa_token=2txe0r_jjhQAAAAA%3ALJa__HZL9COyj9EUpdILZdtnMKLyggfFe7Zpvv0tNze62rLO0CoQHCCJiXfruxUeBLj3YBZ33F8OOv0u">Filling Gaps in Trustworthy Development of AI </a>(Igor is an author on this one)</li><li><a href="https://www.hindawi.com/journals/complexity/2022/8210732/">A Computational Turn in Policy Process Studies: Coevolving Network Dynamics of Policy Change</a></li><li><a href="https://proceedings.neurips.cc/paper_files/paper/2020/file/7e05d6f828574fbc975a896b25bb011e-Paper.pdf">The Smoothed Possibility of Social Choice</a>, an intro in social choice theory and how it overlaps with ML</li><li>Relating to Dan Hendrycks<ul><li><a href="https://arxiv.org/abs/2303.16200">Natural Selection Favors AIs over Humans</a><ul><li>"One easy-to-digest source to highlight what he gets wrong [is] <a href="https://pressbooks.calstate.edu/explorationsbioanth2/chapter/17/">Social and Biopolitical Dimensions of Evolutionary Thinking</a>" -Igor</li></ul></li><li><a href="https://www.aisafetybook.com/">Introduction to AI Safety, Ethics, and Society</a>, recently published textbook</li><li>"<a href="https://arxiv.org/pdf/2306.12001#page=10.19">Source</a> to the section [of this paper] that makes Dan one of my favs from that crowd." -Igor</li><li><a href="https://twitter.com/DanHendrycks/status/1710312043503321141">Twitter post</a> referenced in the episode&lt;...</li></ul></li></ul>]]>
      </itunes:summary>
      <itunes:keywords>AI, machine learning, democracy, AI safety</itunes:keywords>
      <itunes:explicit>Yes</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://krawczuk.eu" img="https://img.transistor.fm/VJHrDED4yyNE9cAwg5Qkc13UxtWcUraWj_NfKbV2HIc/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9jZTE4/YTI3ZTgyYjY4YzFh/OTY0NDkzNmZjZTUz/YjVkYi5qcGVn.jpg">Igor Krawczuk</podcast:person>
      <podcast:chapters url="https://share.transistor.fm/s/b8225038/chapters.json" type="application/json+chapters"/>
    </item>
    <item>
      <title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (3)</title>
      <itunes:episode>18</itunes:episode>
      <podcast:episode>18</podcast:episode>
      <itunes:title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (3)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">0ca91074-f16c-428a-9408-550bfb7ceb4f</guid>
      <link>https://kairos.fm/intoaisafety/e018</link>
      <description>
        <![CDATA[<p>As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.</p><p>As you may have ascertained from the previous two segments of the interview, Dr. Park cofounded <a href="https://www.stakeout.ai">StakeOut.AI</a> along with Harry Luk and one other cofounder whose name has been removed due to requirements of her current position. The non-profit had a simple but important mission: make the adoption of AI technology go well, for humanity, but unfortunately, StakeOut.AI had to dissolve in late February of 2024 because no granter would fund them. Although it certainly is disappointing that the organization is no longer functioning, all three cofounders continue to contribute positively towards improving our world in their current roles.</p><p>If you would like to investigate further into Dr. Park's work, view his <a href="https://scholar.harvard.edu/pspark">website</a>, <a href="https://scholar.google.com/citations?user=5lMAPEoAAAAJ&amp;hl=en">Google Scholar</a>, or follow him on <a href="https://twitter.com/dr_park_phd">Twitter</a></p><p>00:00:54 ❙ Intro<br>00:02:41 ❙ Rapid development<br>00:08:25 ❙ Provable safety, safety factors, &amp; CSAM<br>00:18:50 ❙ Litigation<br>00:23:06 ❙ Open/Closed Source<br>00:38:52 ❙ AIxBio<br>00:47:50 ❙ Scientific rigor in AI<br>00:56:22 ❙ AI deception<br>01:02:45 ❙ No takesies-backsies<br>01:08:22 ❙ StakeOut.AI's start<br>01:12:53 ❙ Sustainability &amp; Agency<br>01:18:21 ❙ "I'm sold, next steps?" -you<br>01:23:53 ❙ Lessons from the amazing Spiderman<br>01:33:15 ❙ "I'm ready to switch careers, next steps?" -you<br>01:40:00 ❙ The most important question<br>01:41:11 ❙ Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://pauseai.info">Pause AI</a></li><li><a href="https://futureoflife.org/wp-content/uploads/2023/11/FLI_Governance_Scorecard_and_Framework.pdf">AI Governance Scorecard</a> (go to Pg. 3)</li><li><a href="https://civitai.com">CIVITAI</a><ul><li><a href="https://www.404media.co/a16z-funded-ai-platform-generated-images-that-could-be-categorized-as-child-pornography-leaked-documents-show/">Article on CIVITAI and CSAM</a></li></ul></li><li><a href="https://www.judiciary.senate.gov/protecting-children-online">Senate Hearing: Protecting Children Online</a><ul><li><a href="https://www.pbs.org/newshour/politics/watch-live-ceos-of-meta-tiktok-x-and-other-social-media-companies-testify-in-senate-hearing">PBS Newshour Coverage</a></li></ul></li><li><a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work</a></li><li>Open Source/Weights/Release/Interpretation<ul><li><a href="https://opensource.org">Open Source Initiative</a><ul><li><a href="https://opensource.org/history">History of the OSI</a></li><li><a href="https://opensource.org/blog/metas-llama-2-license-is-not-open-source">Meta’s LLaMa 2 license is not Open Source</a></li></ul></li><li><a href="https://opensourceconnections.com/blog/2023/07/19/is-llama-2-open-source-no-and-perhaps-we-need-a-new-definition-of-open/">Is Llama 2 open source? No – and perhaps we need a new definition of open…</a></li><li><a href="https://www.apache.org/licenses/LICENSE-2.0">Apache License, Version 2.0</a></li><li><a href="https://www.3blue1brown.com/topics/neural-networks">3Blue1Brown: Neural Networks</a></li><li><a href="https://dl.acm.org/doi/10.1145/3571884.3604316">Opening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generators</a><ul><li>The online <a href="https://opening-up-chatgpt.github.io">table</a></li></ul></li></ul></li><li><a href="https://www.signal.org">Signal</a></li><li><a href="https://huggingface.co/bigscience/bloomz">Bloomz</a> model on HuggingFace</li><li><a href="https://mistral.ai">Mistral</a> website</li><li>NASA Tragedies<ul><li><a href="https://en.wikipedia.org/wiki/Space_Shuttle_Challenger_disaster">Challenger disaster</a> on Wikipedia</li><li><a href="https://en.wikipedia.org/wiki/Space_Shuttle_Columbia_disaster">Columbia disaster</a> on Wikipedia</li></ul></li><li>AIxBio Risk<ul><li><a href="https://www.nature.com/articles/s42256-022-00465-9">Dual use of artificial-intelligence-powered drug discovery</a></li><li><a href="https://arxiv.org/abs/2306.03809">Can large language models democratize access to dual-use biotechnology?</a></li><li><a href="https://www.governance.ai/research-paper/open-sourcing-highly-capable-foundation-models">Open-Sourcing Highly Capable Foundation Models</a> <em>(sadly, I can't rename the article...)</em></li><li><a href="https://1a3orn.com/sub/essays-propaganda-or-science.html">Propaganda or Science: Open Source AI and Bioterrorism Risk</a></li><li><a href="https://ineffectivealtruismblog.com/2024/03/09/exaggerating-the-risks-part-14-biorisk-from-llms/">Exaggerating the risks (Part 15: Biorisk from LLMs)</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will releasing the weights of future large language models grant widespread access to pandemic agents?</a></li><li><a href="https://crfm.stanford.edu/open-fms/">On the Societal Impact of Open Foundation Models</a><ul><li><a href="https://hai.stanford.edu/sites/default/files/2023-12/Governing-Open-Foundation-Models.pdf">Policy brief</a></li></ul></li></ul></li><li><a href="https://www.apartresearch.com">Apart Research</a></li><li><a href="https://www.science.org">Science</a></li><li>Cicero<ul><li><a href="https://www.science.org/doi/10.1126/science.ade9097">Human-level play in the game of Diplomacy by combining language models with strategic reasoning</a></li><li><a href="https://ai.meta.com/research/cicero/">Cicero</a> webpage</li><li><a href="https://arxiv.org/abs/2308.14752">AI Deception: A Survey of Examples, Risks, and Potential Solutions</a></li></ul></li><li><a href="https://demos.co.uk/research/open-sourcing-the-ai-revolution-framing-the-debate-on-open-source-artificial-intelligence-and-regulation/">Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation</a></li><li><a href="https://aisafety.camp">AI Safety Camp</a></li><li><a href="https://www.patreon.com/IntoAISafety">Into AI Safety Patreon</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.</p><p>As you may have ascertained from the previous two segments of the interview, Dr. Park cofounded <a href="https://www.stakeout.ai">StakeOut.AI</a> along with Harry Luk and one other cofounder whose name has been removed due to requirements of her current position. The non-profit had a simple but important mission: make the adoption of AI technology go well, for humanity, but unfortunately, StakeOut.AI had to dissolve in late February of 2024 because no granter would fund them. Although it certainly is disappointing that the organization is no longer functioning, all three cofounders continue to contribute positively towards improving our world in their current roles.</p><p>If you would like to investigate further into Dr. Park's work, view his <a href="https://scholar.harvard.edu/pspark">website</a>, <a href="https://scholar.google.com/citations?user=5lMAPEoAAAAJ&amp;hl=en">Google Scholar</a>, or follow him on <a href="https://twitter.com/dr_park_phd">Twitter</a></p><p>00:00:54 ❙ Intro<br>00:02:41 ❙ Rapid development<br>00:08:25 ❙ Provable safety, safety factors, &amp; CSAM<br>00:18:50 ❙ Litigation<br>00:23:06 ❙ Open/Closed Source<br>00:38:52 ❙ AIxBio<br>00:47:50 ❙ Scientific rigor in AI<br>00:56:22 ❙ AI deception<br>01:02:45 ❙ No takesies-backsies<br>01:08:22 ❙ StakeOut.AI's start<br>01:12:53 ❙ Sustainability &amp; Agency<br>01:18:21 ❙ "I'm sold, next steps?" -you<br>01:23:53 ❙ Lessons from the amazing Spiderman<br>01:33:15 ❙ "I'm ready to switch careers, next steps?" -you<br>01:40:00 ❙ The most important question<br>01:41:11 ❙ Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://pauseai.info">Pause AI</a></li><li><a href="https://futureoflife.org/wp-content/uploads/2023/11/FLI_Governance_Scorecard_and_Framework.pdf">AI Governance Scorecard</a> (go to Pg. 3)</li><li><a href="https://civitai.com">CIVITAI</a><ul><li><a href="https://www.404media.co/a16z-funded-ai-platform-generated-images-that-could-be-categorized-as-child-pornography-leaked-documents-show/">Article on CIVITAI and CSAM</a></li></ul></li><li><a href="https://www.judiciary.senate.gov/protecting-children-online">Senate Hearing: Protecting Children Online</a><ul><li><a href="https://www.pbs.org/newshour/politics/watch-live-ceos-of-meta-tiktok-x-and-other-social-media-companies-testify-in-senate-hearing">PBS Newshour Coverage</a></li></ul></li><li><a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work</a></li><li>Open Source/Weights/Release/Interpretation<ul><li><a href="https://opensource.org">Open Source Initiative</a><ul><li><a href="https://opensource.org/history">History of the OSI</a></li><li><a href="https://opensource.org/blog/metas-llama-2-license-is-not-open-source">Meta’s LLaMa 2 license is not Open Source</a></li></ul></li><li><a href="https://opensourceconnections.com/blog/2023/07/19/is-llama-2-open-source-no-and-perhaps-we-need-a-new-definition-of-open/">Is Llama 2 open source? No – and perhaps we need a new definition of open…</a></li><li><a href="https://www.apache.org/licenses/LICENSE-2.0">Apache License, Version 2.0</a></li><li><a href="https://www.3blue1brown.com/topics/neural-networks">3Blue1Brown: Neural Networks</a></li><li><a href="https://dl.acm.org/doi/10.1145/3571884.3604316">Opening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generators</a><ul><li>The online <a href="https://opening-up-chatgpt.github.io">table</a></li></ul></li></ul></li><li><a href="https://www.signal.org">Signal</a></li><li><a href="https://huggingface.co/bigscience/bloomz">Bloomz</a> model on HuggingFace</li><li><a href="https://mistral.ai">Mistral</a> website</li><li>NASA Tragedies<ul><li><a href="https://en.wikipedia.org/wiki/Space_Shuttle_Challenger_disaster">Challenger disaster</a> on Wikipedia</li><li><a href="https://en.wikipedia.org/wiki/Space_Shuttle_Columbia_disaster">Columbia disaster</a> on Wikipedia</li></ul></li><li>AIxBio Risk<ul><li><a href="https://www.nature.com/articles/s42256-022-00465-9">Dual use of artificial-intelligence-powered drug discovery</a></li><li><a href="https://arxiv.org/abs/2306.03809">Can large language models democratize access to dual-use biotechnology?</a></li><li><a href="https://www.governance.ai/research-paper/open-sourcing-highly-capable-foundation-models">Open-Sourcing Highly Capable Foundation Models</a> <em>(sadly, I can't rename the article...)</em></li><li><a href="https://1a3orn.com/sub/essays-propaganda-or-science.html">Propaganda or Science: Open Source AI and Bioterrorism Risk</a></li><li><a href="https://ineffectivealtruismblog.com/2024/03/09/exaggerating-the-risks-part-14-biorisk-from-llms/">Exaggerating the risks (Part 15: Biorisk from LLMs)</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will releasing the weights of future large language models grant widespread access to pandemic agents?</a></li><li><a href="https://crfm.stanford.edu/open-fms/">On the Societal Impact of Open Foundation Models</a><ul><li><a href="https://hai.stanford.edu/sites/default/files/2023-12/Governing-Open-Foundation-Models.pdf">Policy brief</a></li></ul></li></ul></li><li><a href="https://www.apartresearch.com">Apart Research</a></li><li><a href="https://www.science.org">Science</a></li><li>Cicero<ul><li><a href="https://www.science.org/doi/10.1126/science.ade9097">Human-level play in the game of Diplomacy by combining language models with strategic reasoning</a></li><li><a href="https://ai.meta.com/research/cicero/">Cicero</a> webpage</li><li><a href="https://arxiv.org/abs/2308.14752">AI Deception: A Survey of Examples, Risks, and Potential Solutions</a></li></ul></li><li><a href="https://demos.co.uk/research/open-sourcing-the-ai-revolution-framing-the-debate-on-open-source-artificial-intelligence-and-regulation/">Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation</a></li><li><a href="https://aisafety.camp">AI Safety Camp</a></li><li><a href="https://www.patreon.com/IntoAISafety">Into AI Safety Patreon</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 25 Mar 2024 09:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/8c7eea4d/fac829ef.mp3" length="97914928" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>6120</itunes:duration>
      <itunes:summary>As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.As you may have ascertained from the previous two segments of the interview, Dr. Park cofounded StakeOut.AI along with Harry Luk and one other cofounder whose name has been removed due to requirements of her current position. The non-profit had a simple but important mission: make the adoption of AI technology go well, for humanity, but unfortunately, StakeOut.AI had to dissolve in late February of 2024 because no granter would fund them. Although it certainly is disappointing that the organization is no longer functioning, all three cofounders continue to contribute positively towards improving our world in their current roles.If you would like to investigate further into Dr. Park's work, view his website, Google Scholar, or follow him on Twitter00:00:54 ❙ Intro00:02:41 ❙ Rapid development00:08:25 ❙ Provable safety, safety factors, &amp;amp; CSAM00:18:50 ❙ Litigation00:23:06 ❙ Open/Closed Source00:38:52 ❙ AIxBio00:47:50 ❙ Scientific rigor in AI00:56:22 ❙ AI deception01:02:45 ❙ No takesies-backsies01:08:22 ❙ StakeOut.AI's start01:12:53 ❙ Sustainability &amp;amp; Agency01:18:21 ❙ "I'm sold, next steps?" -you01:23:53 ❙ Lessons from the amazing Spiderman01:33:15 ❙ "I'm ready to switch careers, next steps?" -you01:40:00 ❙ The most important question01:41:11 ❙ OutroLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.StakeOut.AIPause AIAI Governance Scorecard (go to Pg. 3)CIVITAIArticle on CIVITAI and CSAMSenate Hearing: Protecting Children OnlinePBS Newshour CoverageThe Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted WorkOpen Source/Weights/Release/InterpretationOpen Source InitiativeHistory of the OSIMeta’s LLaMa 2 license is not Open SourceIs Llama 2 open source? No – and perhaps we need a new definition of open…Apache License, Version 2.03Blue1Brown: Neural NetworksOpening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generatorsThe online tableSignalBloomz model on HuggingFaceMistral websiteNASA TragediesChallenger disaster on WikipediaColumbia disaster on WikipediaAIxBio RiskDual use of artificial-intelligence-powered drug discoveryCan large language models democratize access to dual-use biotechnology?Open-Sourcing Highly Capable Foundation Models (sadly, I can't rename the article...)Propaganda or Science: Open Source AI and Bioterrorism RiskExaggerating the risks (Part 15: Biorisk from LLMs)Will releasing the weights of future large language models grant widespread access to pandemic agents?On the Societal Impact of Open Foundation ModelsPolicy briefApart ResearchScienceCiceroHuman-level play in the game of Diplomacy by combining language models with strategic reasoningCicero webpageAI Deception: A Survey of Examples, Risks, and Potential SolutionsOpen Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulationAI Safety CampInto AI Safety Patreon</itunes:summary>
      <itunes:subtitle>As always, the best things come in 3s: dimensions, musketeers, pyramids, and... 3 installments of my interview with Dr. Peter Park, an AI Existential Safety Post-doctoral Fellow working with Dr. Max Tegmark at MIT.As you may have ascertained from the prev</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://scholar.harvard.edu/pspark" img="https://img.transistor.fm/fTHAJzIGL_6s2TRpoZXuV1BjZUfMbdgdaI3yJUCCzRI/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9mNjIz/ZmM0ZGJlMDU0NDZj/NDQ4NGVjNjE0YjFm/NDE1MC5qcGc.jpg">Dr. Peter S. Park</podcast:person>
    </item>
    <item>
      <title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (2)</title>
      <itunes:episode>17</itunes:episode>
      <podcast:episode>17</podcast:episode>
      <itunes:title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (2)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">ecb55f43-ac84-4293-b93c-0b9562b236a4</guid>
      <link>https://kairos.fm/intoaisafety/e017</link>
      <description>
        <![CDATA[<p>Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of <a href="https://www.stakeout.ai">StakeOut.AI</a>, a non-profit focused on making AI go well <em>for humans</em>, along with Harry Luk and one other individual, whose name has been removed due to requirements of her current position.</p><p>In addition to the normal links, I wanted to include the links to the petitions that Dr. Park mentions during the podcast. Note that the nonprofit which began these petitions, StakeOut.AI, has been dissolved.<br><a href="https://www.change.org/p/right-ai-laws-to-right-our-future-support-artificial-intelligence-safety-regulations-now">Right AI Laws, to Right Our Future: Support Artificial Intelligence Safety Regulations Now</a><br><a href="https://www.change.org/p/is-deepfake-illegal-not-yet-ban-deepfakes-to-protect-your-family-demand-deepfake-laws">Is Deepfake Illegal? Not Yet! Ban Deepfakes to Protect Your Family &amp; Demand Deepfake Laws</a><br><a href="https://www.change.org/p/ban-superintelligence-stop-ai-driven-human-extinction-risk">Ban Superintelligence: Stop AI-Driven Human Extinction Risk </a></p><p>00:00:54 - Intro<br>00:02:34 - Battleground 1: Copyright<br>00:06:28 - Battleground 2: Moral Critique of AI Collaborationists<br>00:08:15 - Rich Sutton<br>00:20:41 - OpenAI <em>Drama</em><br>00:34:28 - Battleground 3: Contract Negotiations for AI Ban Clauses<br>00:37:57 - Tesla, Autopilot, and FSD<br>00:40:02 - Recycling<br>00:47:40 - Battleground 4: New Laws and Policies<br>00:50:00 - Battleground 5: Whistleblower Protections<br>00:53:07 - Whistleblowing on Microsoft<br>00:54:43 - Andrej Karpathy &amp; Exercises in Empathy<br>01:05:57 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work</a></li><li><a href="https://www.susmangodfrey.com">Susman Godfrey LLP</a></li><li>Rich Sutton<ul><li><a href="http://incompleteideas.net/book/RLbook2020trimmed.pdf">Reinforcement Learning: An Introduction</a> (textbook)</li><li><a href="https://www.youtube.com/watch?v=NgHFMolXs3U">AI Succession</a> (presentation by Rich Sutton)</li><li><a href="https://arxiv.org/abs/2208.11173">The Alberta Plan for AI Research</a></li></ul></li><li>Moore's Law<ul><li><a href="https://www.computerhistory.org/collections/catalog/102770836">The Future of Integrated Electronics</a> (original paper)</li><li><a href="https://www.computerhistory.org/siliconengine/moores-law-predicts-the-future-of-integrated-circuits/">Computer History Museum's entry on Moore's Law</a></li></ul></li><li><a href="https://en.wikipedia.org/wiki/Stochastic_gradient_descent">Stochastic gradient descent (SGD)</a> on Wikipedia</li><li>OpenAI <em>Drama</em><ul><li><a href="https://maxread.substack.com/p/the-interested-normies-guide-to-openai">Max Read's Substack post</a></li><li>Zvi Mowshowitz's Substack series, in order of posting<ul><li><a href="https://thezvi.substack.com/p/openai-facts-from-a-weekend?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Facts from a Weekend</a></li><li><a href="https://thezvi.substack.com/p/openai-the-battle-of-the-board?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: The Battle of the Board</a></li><li><a href="https://thezvi.substack.com/p/openai-altman-returns?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Altman Returns</a></li><li><a href="https://thezvi.substack.com/p/openai-leaks-confirm-the-story?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Leaks Confirm the Story</a> ← best singular post in the series</li><li><a href="https://thezvi.substack.com/p/openai-the-board-expands?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: The Board Expands</a></li></ul></li><li><a href="https://openai.com/blog/openai-announces-new-members-to-board-of-directors">Official OpenAI announcement</a></li></ul></li><li><a href="https://en.wikipedia.org/wiki/Writers_Guild_of_America">WGA</a> on Wikipedia</li><li><a href="https://en.wikipedia.org/wiki/SAG-AFTRA">SAG-AFTRA</a> on Wikipedia</li><li>Tesla's False Advertising<ul><li><a href="https://www.latimes.com/business/story/2023-12-11/tesla-dmv-false-advertising-charges">Tesla's response to the DMV's false-advertising allegations: What took so long?</a></li><li><a href="https://www.caranddriver.com/news/a35785277/tesla-fsd-california-self-driving/">Tesla Tells California DMV that FSD Is Not Capable of Autonomous Driving</a></li><li><a href="https://www.caranddriver.com/features/columns/a35769759/carty-self-driving-cars/">What to Call Full Self-Driving When It Isn't Full Self-Driving?</a></li><li><a href="https://www.cnbc.com/2022/03/15/tesla-fired-employee-who-posted-fsd-beta-videos-as-ai-addict-on-youtube.html">Tesla fired an employee after he posted driverless tech reviews on YouTube</a></li><li><a href="https://www.tesla.com/support/autopilot">Tesla's page</a> on Autopilot and Full Self-Driving</li></ul></li><li>Recycling<ul><li><a href="https://www.longmontleader.com/regional-news/boulder-county-recycling-center-stockpiles-accurately-sorted-recyclable-materials-2385879">Boulder County Recycling Center Stockpiles Accurately Sorted Recyclable Materials</a></li><li><a href="https://boulderweekly.com/news/out-of-sight-out-of-mind/">Out of sight, out of mind</a></li><li><a href="https://ecocycle.org/content/uploads/2022/06/2024-Recycling-Guidelines-for-Boulder-County.pdf">Boulder Eco-Cycle Recycling Guidelines</a></li></ul></li><li><a href="https://arxiv.org/abs/2310.06009">Divide-and-Conquer Dynamics in AI-Driven Disempowerment</a></li><li>Microsoft Whistleblower<ul><li><a href="https://www.axios.com/2024/03/08/ai-whistleblowers-microsoft-copilot-designer-google-gemini">Whistleblowers call out AI's flaws</a></li><li><a href="https://www.linkedin.com/feed/update/urn:li:activity:7171135079702753280/">Shane's LinkedIn post</a><ul><li><a href="https://media.licdn.com/dms/document/media/D561FAQFfYnpLbIn2Xg/feedshare-document-pdf-analyzed/0/1709731400675?e=1711584000&amp;v=beta&amp;t=W3CLljNWJ8YCQSEuMCR9bt7UoIxsGo24Epj9GddfN8U">Letters sent by Jones</a></li></ul></li></ul></li><li><a href="https://twitter.com/karpathy/status/1757600075281547344">Karpathy announces departure from OpenAI</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of <a href="https://www.stakeout.ai">StakeOut.AI</a>, a non-profit focused on making AI go well <em>for humans</em>, along with Harry Luk and one other individual, whose name has been removed due to requirements of her current position.</p><p>In addition to the normal links, I wanted to include the links to the petitions that Dr. Park mentions during the podcast. Note that the nonprofit which began these petitions, StakeOut.AI, has been dissolved.<br><a href="https://www.change.org/p/right-ai-laws-to-right-our-future-support-artificial-intelligence-safety-regulations-now">Right AI Laws, to Right Our Future: Support Artificial Intelligence Safety Regulations Now</a><br><a href="https://www.change.org/p/is-deepfake-illegal-not-yet-ban-deepfakes-to-protect-your-family-demand-deepfake-laws">Is Deepfake Illegal? Not Yet! Ban Deepfakes to Protect Your Family &amp; Demand Deepfake Laws</a><br><a href="https://www.change.org/p/ban-superintelligence-stop-ai-driven-human-extinction-risk">Ban Superintelligence: Stop AI-Driven Human Extinction Risk </a></p><p>00:00:54 - Intro<br>00:02:34 - Battleground 1: Copyright<br>00:06:28 - Battleground 2: Moral Critique of AI Collaborationists<br>00:08:15 - Rich Sutton<br>00:20:41 - OpenAI <em>Drama</em><br>00:34:28 - Battleground 3: Contract Negotiations for AI Ban Clauses<br>00:37:57 - Tesla, Autopilot, and FSD<br>00:40:02 - Recycling<br>00:47:40 - Battleground 4: New Laws and Policies<br>00:50:00 - Battleground 5: Whistleblower Protections<br>00:53:07 - Whistleblowing on Microsoft<br>00:54:43 - Andrej Karpathy &amp; Exercises in Empathy<br>01:05:57 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://www.nytimes.com/2023/12/27/business/media/new-york-times-open-ai-microsoft-lawsuit.html">The Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted Work</a></li><li><a href="https://www.susmangodfrey.com">Susman Godfrey LLP</a></li><li>Rich Sutton<ul><li><a href="http://incompleteideas.net/book/RLbook2020trimmed.pdf">Reinforcement Learning: An Introduction</a> (textbook)</li><li><a href="https://www.youtube.com/watch?v=NgHFMolXs3U">AI Succession</a> (presentation by Rich Sutton)</li><li><a href="https://arxiv.org/abs/2208.11173">The Alberta Plan for AI Research</a></li></ul></li><li>Moore's Law<ul><li><a href="https://www.computerhistory.org/collections/catalog/102770836">The Future of Integrated Electronics</a> (original paper)</li><li><a href="https://www.computerhistory.org/siliconengine/moores-law-predicts-the-future-of-integrated-circuits/">Computer History Museum's entry on Moore's Law</a></li></ul></li><li><a href="https://en.wikipedia.org/wiki/Stochastic_gradient_descent">Stochastic gradient descent (SGD)</a> on Wikipedia</li><li>OpenAI <em>Drama</em><ul><li><a href="https://maxread.substack.com/p/the-interested-normies-guide-to-openai">Max Read's Substack post</a></li><li>Zvi Mowshowitz's Substack series, in order of posting<ul><li><a href="https://thezvi.substack.com/p/openai-facts-from-a-weekend?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Facts from a Weekend</a></li><li><a href="https://thezvi.substack.com/p/openai-the-battle-of-the-board?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: The Battle of the Board</a></li><li><a href="https://thezvi.substack.com/p/openai-altman-returns?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Altman Returns</a></li><li><a href="https://thezvi.substack.com/p/openai-leaks-confirm-the-story?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: Leaks Confirm the Story</a> ← best singular post in the series</li><li><a href="https://thezvi.substack.com/p/openai-the-board-expands?utm_source=%2Fsearch%2FOpenAI&amp;utm_medium=reader2">OpenAI: The Board Expands</a></li></ul></li><li><a href="https://openai.com/blog/openai-announces-new-members-to-board-of-directors">Official OpenAI announcement</a></li></ul></li><li><a href="https://en.wikipedia.org/wiki/Writers_Guild_of_America">WGA</a> on Wikipedia</li><li><a href="https://en.wikipedia.org/wiki/SAG-AFTRA">SAG-AFTRA</a> on Wikipedia</li><li>Tesla's False Advertising<ul><li><a href="https://www.latimes.com/business/story/2023-12-11/tesla-dmv-false-advertising-charges">Tesla's response to the DMV's false-advertising allegations: What took so long?</a></li><li><a href="https://www.caranddriver.com/news/a35785277/tesla-fsd-california-self-driving/">Tesla Tells California DMV that FSD Is Not Capable of Autonomous Driving</a></li><li><a href="https://www.caranddriver.com/features/columns/a35769759/carty-self-driving-cars/">What to Call Full Self-Driving When It Isn't Full Self-Driving?</a></li><li><a href="https://www.cnbc.com/2022/03/15/tesla-fired-employee-who-posted-fsd-beta-videos-as-ai-addict-on-youtube.html">Tesla fired an employee after he posted driverless tech reviews on YouTube</a></li><li><a href="https://www.tesla.com/support/autopilot">Tesla's page</a> on Autopilot and Full Self-Driving</li></ul></li><li>Recycling<ul><li><a href="https://www.longmontleader.com/regional-news/boulder-county-recycling-center-stockpiles-accurately-sorted-recyclable-materials-2385879">Boulder County Recycling Center Stockpiles Accurately Sorted Recyclable Materials</a></li><li><a href="https://boulderweekly.com/news/out-of-sight-out-of-mind/">Out of sight, out of mind</a></li><li><a href="https://ecocycle.org/content/uploads/2022/06/2024-Recycling-Guidelines-for-Boulder-County.pdf">Boulder Eco-Cycle Recycling Guidelines</a></li></ul></li><li><a href="https://arxiv.org/abs/2310.06009">Divide-and-Conquer Dynamics in AI-Driven Disempowerment</a></li><li>Microsoft Whistleblower<ul><li><a href="https://www.axios.com/2024/03/08/ai-whistleblowers-microsoft-copilot-designer-google-gemini">Whistleblowers call out AI's flaws</a></li><li><a href="https://www.linkedin.com/feed/update/urn:li:activity:7171135079702753280/">Shane's LinkedIn post</a><ul><li><a href="https://media.licdn.com/dms/document/media/D561FAQFfYnpLbIn2Xg/feedshare-document-pdf-analyzed/0/1709731400675?e=1711584000&amp;v=beta&amp;t=W3CLljNWJ8YCQSEuMCR9bt7UoIxsGo24Epj9GddfN8U">Letters sent by Jones</a></li></ul></li></ul></li><li><a href="https://twitter.com/karpathy/status/1757600075281547344">Karpathy announces departure from OpenAI</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 18 Mar 2024 09:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/58dfc367/2d88b642.mp3" length="63715835" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>3983</itunes:duration>
      <itunes:summary>Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of StakeOut.AI, a non-profit focused on making AI go well for humans, along with Harry Luk and one other individual, whose name has been removed due to requirements of her current position.In addition to the normal links, I wanted to include the links to the petitions that Dr. Park mentions during the podcast. Note that the nonprofit which began these petitions, StakeOut.AI, has been dissolved.Right AI Laws, to Right Our Future: Support Artificial Intelligence Safety Regulations NowIs Deepfake Illegal? Not Yet! Ban Deepfakes to Protect Your Family &amp;amp; Demand Deepfake LawsBan Superintelligence: Stop AI-Driven Human Extinction Risk
00:00:54 - Intro00:02:34 - Battleground 1: Copyright00:06:28 - Battleground 2: Moral Critique of AI Collaborationists00:08:15 - Rich Sutton00:20:41 - OpenAI Drama00:34:28 - Battleground 3: Contract Negotiations for AI Ban Clauses00:37:57 - Tesla, Autopilot, and FSD00:40:02 - Recycling00:47:40 - Battleground 4: New Laws and Policies00:50:00 - Battleground 5: Whistleblower Protections00:53:07 - Whistleblowing on Microsoft00:54:43 - Andrej Karpathy &amp;amp; Exercises in Empathy01:05:57 - OutroLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.StakeOut.AIThe Times Sues OpenAI and Microsoft Over A.I. Use of Copyrighted WorkSusman Godfrey LLPRich SuttonReinforcement Learning: An Introduction (textbook)AI Succession (presentation by Rich Sutton)The Alberta Plan for AI Research Moore's LawThe Future of Integrated Electronics (original paper)Computer History Museum's entry on Moore's LawStochastic gradient descent (SGD) on WikipediaOpenAI DramaMax Read's Substack postZvi Mowshowitz's Substack series, in order of postingOpenAI: Facts from a WeekendOpenAI: The Battle of the BoardOpenAI: Altman ReturnsOpenAI: Leaks Confirm the Story ← best singular post in the seriesOpenAI: The Board ExpandsOfficial OpenAI announcementWGA on WikipediaSAG-AFTRA on WikipediaTesla's False AdvertisingTesla's response to the DMV's false-advertising allegations: What took so long?Tesla Tells California DMV that FSD Is Not Capable of Autonomous DrivingWhat to Call Full Self-Driving When It Isn't Full Self-Driving?Tesla fired an employee after he posted driverless tech reviews on YouTubeTesla's page on Autopilot and Full Self-DrivingRecyclingBoulder County Recycling Center Stockpiles Accurately Sorted Recyclable MaterialsOut of sight, out of mindBoulder Eco-Cycle Recycling GuidelinesDivide-and-Conquer Dynamics in AI-Driven DisempowermentMicrosoft WhistleblowerWhistleblowers call out AI's flawsShane's LinkedIn postLetters sent by JonesKarpathy announces departure from OpenAI</itunes:summary>
      <itunes:subtitle>Join me for round 2 with Dr. Peter Park, an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. Dr. Park was a cofounder of StakeOut.AI, a non-profit focused on making AI go well for humans, along with Harry Luk and one other in</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://scholar.harvard.edu/pspark" img="https://img.transistor.fm/fTHAJzIGL_6s2TRpoZXuV1BjZUfMbdgdaI3yJUCCzRI/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9mNjIz/ZmM0ZGJlMDU0NDZj/NDQ4NGVjNjE0YjFm/NDE1MC5qcGc.jpg">Dr. Peter S. Park</podcast:person>
    </item>
    <item>
      <title>MINISODE: Restructure Vol. 2</title>
      <itunes:episode>16</itunes:episode>
      <podcast:episode>16</podcast:episode>
      <itunes:title>MINISODE: Restructure Vol. 2</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">6dc7d6e0-46ac-49ce-a60f-fe7c86e3f648</guid>
      <link>https://kairos.fm/intoaisafety/e016</link>
      <description>
        <![CDATA[<p>UPDATE: Contrary to what I say in this episode, I won't be removing <em>any</em> episodes that are already published from the podcast RSS feed.</p><p>After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the podcast towards accessible content regarding "AI" instead of the show's original focus. I will still be releasing what I am calling research ride-along content to my <a href="https://www.patreon.com/IntoAISafety">Patreon</a>, but the show's feed will consist only of content that I aim to make as accessible as possible.</p><p>00:35 - TL;DL<br>01:12 - Advice from Pete<br>03:10 - My personal goal<br>05:39 - Reflection on refining my goal<br>09:08 - Looking forward (logistics</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>UPDATE: Contrary to what I say in this episode, I won't be removing <em>any</em> episodes that are already published from the podcast RSS feed.</p><p>After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the podcast towards accessible content regarding "AI" instead of the show's original focus. I will still be releasing what I am calling research ride-along content to my <a href="https://www.patreon.com/IntoAISafety">Patreon</a>, but the show's feed will consist only of content that I aim to make as accessible as possible.</p><p>00:35 - TL;DL<br>01:12 - Advice from Pete<br>03:10 - My personal goal<br>05:39 - Reflection on refining my goal<br>09:08 - Looking forward (logistics</p>]]>
      </content:encoded>
      <pubDate>Mon, 11 Mar 2024 09:00:00 -0600</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/fa6df270/5048eeef.mp3" length="12611516" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>789</itunes:duration>
      <itunes:summary>UPDATE: Contrary to what I say in this episode, I won't be removing any episodes that are already published from the podcast RSS feed.
After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the podcast towards accessible content regarding "AI" instead of the show's original focus. I will still be releasing what I am calling research ride-along content to my Patreon, but the show's feed will consist only of content that I aim to make as accessible as possible.
00:35 - TL;DL01:12 - Advice from Pete03:10 - My personal goal05:39 - Reflection on refining my goal09:08 - Looking forward (logistics</itunes:summary>
      <itunes:subtitle>UPDATE: Contrary to what I say in this episode, I won't be removing any episodes that are already published from the podcast RSS feed.
After getting some advice and reflecting more on my own personal goals, I have decided to shift the direction of the po</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (1)</title>
      <itunes:episode>15</itunes:episode>
      <podcast:episode>15</podcast:episode>
      <itunes:title>INTERVIEW: StakeOut.AI w/ Dr. Peter Park (1)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">cd0206b8-6690-44c7-97e3-240c2b082096</guid>
      <link>https://kairos.fm/intoaisafety/e015</link>
      <description>
        <![CDATA[<p>Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded <a href="https://www.stakeout.ai/">⁠StakeOut.AI</a>, a non-profit focused on making AI go well <em>for humans</em>.</p><p>00:54 - Intro<br>03:15 - Dr. Park, x-risk, and AGI<br>08:55 - StakeOut.AI<br>12:05 - Governance scorecard<br>19:34 - Hollywood webinar<br>22:02 - Regulations.gov comments<br>23:48 - Open letters <br>26:15 - EU AI Act<br>35:07 - Effective accelerationism<br>40:50 - Divide and conquer dynamics<br>45:40 - AI "art"<br>53:09 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://futureoflife.org/wp-content/uploads/2023/11/FLI_Governance_Scorecard_and_Framework.pdf">AI Governance Scorecard</a> (go to Pg. 3)</li><li><a href="https://pauseai.info">Pause AI</a></li><li><a href="https://www.regulations.gov">Regulations.gov</a><br> <ul><li><a href="https://www.regulations.gov/comment/COLC-2023-0006-10077">USCO StakeOut.AI Comment</a></li></ul></li><li> <ul><li><a href="https://www.regulations.gov/comment/OMB-2023-0020-0170">OMB StakeOut.AI Comment</a></li></ul></li><li> </li><li><a href="https://aitreaty.org">AI Treaty open letter</a></li><li><a href="https://taisc.org">TAISC</a></li><li><a href="https://crfm.stanford.edu/2023/03/13/alpaca.html">Alpaca: A Strong, Replicable Instruction-Following Model</a></li><li>References on EU AI Act and Cedric O<br> <ul><li><a href="https://twitter.com/cedric_o/status/1728724005459235052">Tweet from Cedric O</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/eu-policymakers-enter-the-last-mile-for-artificial-intelligence-rulebook/">EU policymakers enter the last mile for Artificial Intelligence rulebook</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/ai-act-eu-parliaments-legal-office-gives-damning-opinion-on-high-risk-classification-filters/">AI Act: EU Parliament’s legal office gives damning opinion on high-risk classification ‘filters’</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/eus-ai-act-negotiations-hit-the-brakes-over-foundation-models/">EU’s AI Act negotiations hit the brakes over foundation models</a></li></ul></li><li> <ul><li><a href="https://www.foundation-models.eu">The EU AI Act needs Foundation Model Regulation</a></li></ul></li><li> <ul><li><a href="https://verfassungsblog.de/bigtechs-efforts-to-derail-the-ai-act/">BigTech’s Efforts to Derail the AI Act</a></li></ul></li><li> </li><li><a href="https://demos.co.uk/research/open-sourcing-the-ai-revolution-framing-the-debate-on-open-source-artificial-intelligence-and-regulation/">Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation</a></li><li><a href="https://arxiv.org/abs/2310.06009">Divide-and-Conquer Dynamics in AI-Driven Disempowerment</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded <a href="https://www.stakeout.ai/">⁠StakeOut.AI</a>, a non-profit focused on making AI go well <em>for humans</em>.</p><p>00:54 - Intro<br>03:15 - Dr. Park, x-risk, and AGI<br>08:55 - StakeOut.AI<br>12:05 - Governance scorecard<br>19:34 - Hollywood webinar<br>22:02 - Regulations.gov comments<br>23:48 - Open letters <br>26:15 - EU AI Act<br>35:07 - Effective accelerationism<br>40:50 - Divide and conquer dynamics<br>45:40 - AI "art"<br>53:09 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.stakeout.ai">StakeOut.AI</a></li><li><a href="https://futureoflife.org/wp-content/uploads/2023/11/FLI_Governance_Scorecard_and_Framework.pdf">AI Governance Scorecard</a> (go to Pg. 3)</li><li><a href="https://pauseai.info">Pause AI</a></li><li><a href="https://www.regulations.gov">Regulations.gov</a><br> <ul><li><a href="https://www.regulations.gov/comment/COLC-2023-0006-10077">USCO StakeOut.AI Comment</a></li></ul></li><li> <ul><li><a href="https://www.regulations.gov/comment/OMB-2023-0020-0170">OMB StakeOut.AI Comment</a></li></ul></li><li> </li><li><a href="https://aitreaty.org">AI Treaty open letter</a></li><li><a href="https://taisc.org">TAISC</a></li><li><a href="https://crfm.stanford.edu/2023/03/13/alpaca.html">Alpaca: A Strong, Replicable Instruction-Following Model</a></li><li>References on EU AI Act and Cedric O<br> <ul><li><a href="https://twitter.com/cedric_o/status/1728724005459235052">Tweet from Cedric O</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/eu-policymakers-enter-the-last-mile-for-artificial-intelligence-rulebook/">EU policymakers enter the last mile for Artificial Intelligence rulebook</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/ai-act-eu-parliaments-legal-office-gives-damning-opinion-on-high-risk-classification-filters/">AI Act: EU Parliament’s legal office gives damning opinion on high-risk classification ‘filters’</a></li></ul></li><li> <ul><li><a href="https://www.euractiv.com/section/artificial-intelligence/news/eus-ai-act-negotiations-hit-the-brakes-over-foundation-models/">EU’s AI Act negotiations hit the brakes over foundation models</a></li></ul></li><li> <ul><li><a href="https://www.foundation-models.eu">The EU AI Act needs Foundation Model Regulation</a></li></ul></li><li> <ul><li><a href="https://verfassungsblog.de/bigtechs-efforts-to-derail-the-ai-act/">BigTech’s Efforts to Derail the AI Act</a></li></ul></li><li> </li><li><a href="https://demos.co.uk/research/open-sourcing-the-ai-revolution-framing-the-debate-on-open-source-artificial-intelligence-and-regulation/">Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation</a></li><li><a href="https://arxiv.org/abs/2310.06009">Divide-and-Conquer Dynamics in AI-Driven Disempowerment</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 04 Mar 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/81a6fb5c/cb1fde4e.mp3" length="52004201" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>3251</itunes:duration>
      <itunes:summary>Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded 
⁠StakeOut.AI, a non-profit focused on making AI go well for humans.
00:54 - Intro03:15 - Dr. Park, x-risk, and AGI08:55 - StakeOut.AI12:05 - Governance scorecard19:34 - Hollywood webinar22:02 - Regulations.gov comments23:48 - Open letters 26:15 - EU AI Act35:07 - Effective accelerationism40:50 - Divide and conquer dynamics45:40 - AI "art"53:09 - Outro
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.

 StakeOut.AI
 AI Governance Scorecard (go to Pg. 3)
 Pause AI
 Regulations.gov
  USCO StakeOut.AI Comment
  OMB StakeOut.AI Comment


  AI Treaty open letter
  TAISC
  Alpaca: A Strong, Replicable Instruction-Following Model
  References on EU AI Act and Cedric O
  Tweet from Cedric O
  EU policymakers enter the last mile for Artificial Intelligence rulebook
  AI Act: EU Parliament’s legal office gives damning opinion on high-risk classification ‘filters’
  EU’s AI Act negotiations hit the brakes over foundation models
  The EU AI Act needs Foundation Model Regulation
  BigTech’s Efforts to Derail the AI Act


  Open Sourcing the AI Revolution: Framing the debate on open source, artificial intelligence and regulation
  Divide-and-Conquer Dynamics in AI-Driven Disempowerment</itunes:summary>
      <itunes:subtitle>Dr. Peter Park is an AI Existential Safety Postdoctoral Fellow working with Dr. Max Tegmark at MIT. In conjunction with Harry Luk and one other cofounder, he founded 
⁠StakeOut.AI, a non-profit focused on making AI go well for humans.
00:54 - Intro03:15</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
      <podcast:person role="Guest" href="https://scholar.harvard.edu/pspark" img="https://img.transistor.fm/fTHAJzIGL_6s2TRpoZXuV1BjZUfMbdgdaI3yJUCCzRI/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9mNjIz/ZmM0ZGJlMDU0NDZj/NDQ4NGVjNjE0YjFm/NDE1MC5qcGc.jpg">Dr. Peter S. Park</podcast:person>
    </item>
    <item>
      <title>MINISODE: "LLMs, a Survey"</title>
      <itunes:episode>14</itunes:episode>
      <podcast:episode>14</podcast:episode>
      <itunes:title>MINISODE: "LLMs, a Survey"</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">2d7e4db8-5798-4148-9e19-53a5143c4919</guid>
      <link>https://kairos.fm/intoaisafety/e014</link>
      <description>
        <![CDATA[<p>Take a trip with me through the paper <a href="https://arxiv.org/abs/2402.06196">Large Language Models, A Survey</a>, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety <a href="https://into-ai-safety.github.io">podcast website</a>.</p><p>00:36 - Intro and authors<br>01:50 - My takes and paper structure<br>04:40 - Getting to LLMs<br>07:27 - Defining LLMs &amp; emergence<br>12:12 - Overview of PLMs<br>15:00 - How LLMs are built<br>18:52 - Limitations if LLMs<br>23:06 - Uses of LLMs<br>25:16 - Evaluations and Benchmarks<br>28:11 - Challenges and future directions<br>29:21 - Recap &amp; outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://arxiv.org/abs/2402.06196">Large Language Models, A Survey</a></li><li><a href="https://www.linkedin.com/posts/meysam-ac_i-am-delighted-to-share-that-our-most-recent-activity-7162768857827377152-wiLu/?utm_source=share&amp;utm_medium=member_desktop">Meysam's LinkedIn Post</a></li><li>Claude E. Shannon<ul><li><a href="https://dspace.mit.edu/handle/1721.1/11173">A symbolic analysis of relay and switching circuits</a> (Master's Thesis)</li><li><a href="https://ieeexplore.ieee.org/document/6769090">Communication theory of secrecy systems</a></li><li><a href="https://ieeexplore.ieee.org/document/6773024">A mathematical theory of communication</a></li><li><a href="https://ieeexplore.ieee.org/document/6773263">Prediction and entropy of printed English</a></li></ul></li><li><a href="https://bounded-regret.ghost.io/future-ml-systems-will-be-qualitatively-different/">Future ML Systems Will Be Qualitatively Different</a></li><li><a href="https://www.science.org/doi/10.1126/science.177.4047.393?ref=bounded-regret.ghost.io">More Is Different</a></li><li><a href="https://arxiv.org/abs/2401.05566">Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training</a></li><li><a href="https://arxiv.org/abs/2304.15004">Are Emergent Abilities of Large Language Models a Mirage?</a></li><li><a href="https://arxiv.org/abs/2309.01809">Are Emergent Abilities of Large Language Models just In-Context Learning?</a></li><li><a href="https://arxiv.org/abs/1706.03762">Attention is all you need</a></li><li><a href="https://arxiv.org/abs/2305.18290">Direct Preference Optimization: Your Language Model is Secretly a Reward Model</a></li><li><a href="https://arxiv.org/abs/2402.01306">KTO: Model Alignment as Prospect Theoretic Optimization</a></li><li><a href="https://www2.stat.duke.edu/~scs/Courses/Stat376/Papers/TemperAnneal/KirkpatrickAnnealScience1983.pdf">Optimization by Simulated Annealing</a></li><li><a href="https://openai.com/blog/memory-and-new-controls-for-chatgpt">Memory and new controls for ChatGPT</a></li><li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4515540/">Hallucinations and related concepts—their conceptual background</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Take a trip with me through the paper <a href="https://arxiv.org/abs/2402.06196">Large Language Models, A Survey</a>, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety <a href="https://into-ai-safety.github.io">podcast website</a>.</p><p>00:36 - Intro and authors<br>01:50 - My takes and paper structure<br>04:40 - Getting to LLMs<br>07:27 - Defining LLMs &amp; emergence<br>12:12 - Overview of PLMs<br>15:00 - How LLMs are built<br>18:52 - Limitations if LLMs<br>23:06 - Uses of LLMs<br>25:16 - Evaluations and Benchmarks<br>28:11 - Challenges and future directions<br>29:21 - Recap &amp; outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://arxiv.org/abs/2402.06196">Large Language Models, A Survey</a></li><li><a href="https://www.linkedin.com/posts/meysam-ac_i-am-delighted-to-share-that-our-most-recent-activity-7162768857827377152-wiLu/?utm_source=share&amp;utm_medium=member_desktop">Meysam's LinkedIn Post</a></li><li>Claude E. Shannon<ul><li><a href="https://dspace.mit.edu/handle/1721.1/11173">A symbolic analysis of relay and switching circuits</a> (Master's Thesis)</li><li><a href="https://ieeexplore.ieee.org/document/6769090">Communication theory of secrecy systems</a></li><li><a href="https://ieeexplore.ieee.org/document/6773024">A mathematical theory of communication</a></li><li><a href="https://ieeexplore.ieee.org/document/6773263">Prediction and entropy of printed English</a></li></ul></li><li><a href="https://bounded-regret.ghost.io/future-ml-systems-will-be-qualitatively-different/">Future ML Systems Will Be Qualitatively Different</a></li><li><a href="https://www.science.org/doi/10.1126/science.177.4047.393?ref=bounded-regret.ghost.io">More Is Different</a></li><li><a href="https://arxiv.org/abs/2401.05566">Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training</a></li><li><a href="https://arxiv.org/abs/2304.15004">Are Emergent Abilities of Large Language Models a Mirage?</a></li><li><a href="https://arxiv.org/abs/2309.01809">Are Emergent Abilities of Large Language Models just In-Context Learning?</a></li><li><a href="https://arxiv.org/abs/1706.03762">Attention is all you need</a></li><li><a href="https://arxiv.org/abs/2305.18290">Direct Preference Optimization: Your Language Model is Secretly a Reward Model</a></li><li><a href="https://arxiv.org/abs/2402.01306">KTO: Model Alignment as Prospect Theoretic Optimization</a></li><li><a href="https://www2.stat.duke.edu/~scs/Courses/Stat376/Papers/TemperAnneal/KirkpatrickAnnealScience1983.pdf">Optimization by Simulated Annealing</a></li><li><a href="https://openai.com/blog/memory-and-new-controls-for-chatgpt">Memory and new controls for ChatGPT</a></li><li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4515540/">Hallucinations and related concepts—their conceptual background</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 26 Feb 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/24982632/d6945322.mp3" length="29667591" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>1855</itunes:duration>
      <itunes:summary>Take a trip with me through the paper Large Language Models, A Survey, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety podcast website.
00:36 - Intro and authors01:50 - My takes and paper structure04:40 - Getting to LLMs07:27 - Defining LLMs &amp;amp; emergence12:12 - Overview of PLMs15:00 - How LLMs are built18:52 - Limitations if LLMs23:06 - Uses of LLMs25:16 - Evaluations and Benchmarks28:11 - Challenges and future directions29:21 - Recap &amp;amp; outro
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.Large Language Models, A SurveyMeysam's LinkedIn PostClaude E. ShannonA symbolic analysis of relay and switching circuits (Master's Thesis)Communication theory of secrecy systemsA mathematical theory of communicationPrediction and entropy of printed EnglishFuture ML Systems Will Be Qualitatively DifferentMore Is DifferentSleeper Agents: Training Deceptive LLMs that Persist Through Safety TrainingAre Emergent Abilities of Large Language Models a Mirage?Are Emergent Abilities of Large Language Models just In-Context Learning?Attention is all you needDirect Preference Optimization: Your Language Model is Secretly a Reward ModelKTO: Model Alignment as Prospect Theoretic OptimizationOptimization by Simulated AnnealingMemory and new controls for ChatGPTHallucinations and related concepts—their conceptual background</itunes:summary>
      <itunes:subtitle>Take a trip with me through the paper Large Language Models, A Survey, published on February 9th of 2024. All figures and tables mentioned throughout the episode can be found on the Into AI Safety podcast website.
00:36 - Intro and authors01:50 - My take</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>FEEDBACK: Applying for Funding w/ Esben Kran</title>
      <itunes:episode>13</itunes:episode>
      <podcast:episode>13</podcast:episode>
      <itunes:title>FEEDBACK: Applying for Funding w/ Esben Kran</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">9c9288c4-d48b-48fa-aec6-b11cc5803619</guid>
      <link>https://kairos.fm/intoaisafety/e013</link>
      <description>
        <![CDATA[<p>Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when applying for funding in the future.</p><p>Head over to Apart Research's <a href="https://apartresearch.com">website</a> to check out their work, or the Alignment Jam <a href="https://alignmentjam.com">website</a> for information on upcoming hackathons.</p><p>A doc-capsule of the application at the time of this recording can be found <a href="https://docs.google.com/document/d/1ofk5VLvNeJytd5Rpx18zui-JfV900DZv5nZShZcrgfY/edit?usp=sharing">at this link</a>.</p><p>01:38 - Interview starts<br>05:41 - Proposal<br>11:00 - Personal statement<br>14:00 - Budget<br>21:12 - CV<br>22:45 - Application questions<br>34:06 - Funding questions<br>44:25 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.openphilanthropy.org/research/ai-governance-talent-profiles-wed-like-to-see/">AI governance talent profiles we’d like to see</a></li><li><a href="https://alignmentjam.com/jam/governance">The AI Governance Research Sprint</a></li><li><a href="https://www.openphilanthropy.org/research/reasoning-transparency/">Reasoning Transparency</a></li><li>Places to look for funding<ul><li>Open Philanthropy's <a href="https://www.openphilanthropy.org/career-development-and-transition-funding/">Career development and transition funding</a></li><li><a href="https://funds.effectivealtruism.org/funds/far-future">Long-Term Future Fund</a></li><li><a href="https://manifund.org">Manifund</a></li></ul></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when applying for funding in the future.</p><p>Head over to Apart Research's <a href="https://apartresearch.com">website</a> to check out their work, or the Alignment Jam <a href="https://alignmentjam.com">website</a> for information on upcoming hackathons.</p><p>A doc-capsule of the application at the time of this recording can be found <a href="https://docs.google.com/document/d/1ofk5VLvNeJytd5Rpx18zui-JfV900DZv5nZShZcrgfY/edit?usp=sharing">at this link</a>.</p><p>01:38 - Interview starts<br>05:41 - Proposal<br>11:00 - Personal statement<br>14:00 - Budget<br>21:12 - CV<br>22:45 - Application questions<br>34:06 - Funding questions<br>44:25 - Outro</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.openphilanthropy.org/research/ai-governance-talent-profiles-wed-like-to-see/">AI governance talent profiles we’d like to see</a></li><li><a href="https://alignmentjam.com/jam/governance">The AI Governance Research Sprint</a></li><li><a href="https://www.openphilanthropy.org/research/reasoning-transparency/">Reasoning Transparency</a></li><li>Places to look for funding<ul><li>Open Philanthropy's <a href="https://www.openphilanthropy.org/career-development-and-transition-funding/">Career development and transition funding</a></li><li><a href="https://funds.effectivealtruism.org/funds/far-future">Long-Term Future Fund</a></li><li><a href="https://manifund.org">Manifund</a></li></ul></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 19 Feb 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/aa6ca492/29c49e43.mp3" length="43395914" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>2713</itunes:duration>
      <itunes:summary>Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when applying for funding in the future.Head over to Apart Research's website to check out their work, or the Alignment Jam website for information on upcoming hackathons.A doc-capsule of the application at the time of this recording can be found at this link.01:38 - Interview starts05:41 - Proposal11:00 - Personal statement14:00 - Budget21:12 - CV22:45 - Application questions34:06 - Funding questions44:25 - OutroLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.AI governance talent profiles we’d like to seeThe AI Governance Research SprintReasoning TransparencyPlaces to look for fundingOpen Philanthropy's Career development and transition fundingLong-Term Future FundManifund</itunes:summary>
      <itunes:subtitle>Esben reviews an application that I would soon submit for Open Philanthropy's Career Transitition Funding opportunity. Although I didn't end up receiving the funding, I do think that this episode can be a valuable resource for both others and myself when </itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Reading a Research Paper</title>
      <itunes:episode>12</itunes:episode>
      <podcast:episode>12</podcast:episode>
      <itunes:title>MINISODE: Reading a Research Paper</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">37fd4eb5-ba2f-419e-aa24-9a1425a69051</guid>
      <link>https://kairos.fm/intoaisafety/e012</link>
      <description>
        <![CDATA[<p>Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice.</p><p>00:56 - Anatomy of a paper<br>02:38 - Most common advice<br>05:24 - Reading sparsity and path<br>07:30 - Notes and motivation</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7392212/">Ten simple rules for reading a scientific paper</a></li><li>Best sources I found<ul><li><a href="https://subjectguides.york.ac.uk/critical/articles">Let's get critical: Reading academic articles</a></li><li><a href="https://www.scientifica.uk.com/neurowire/gradhacks-a-guide-to-reading-research-papers">#GradHacks: A guide to reading research papers</a></li><li><a href="https://www.lib.purdue.edu/sites/default/files/libraries/engr/Tutorials/Newest%20Scientific%20Paper.pdf">How to read a scientific paper</a> (presentation)</li></ul></li><li>Some more sources<ul><li><a href="https://www.owlnet.rice.edu/~cainproj/courses/HowToReadSciArticle.pdf">How to read a scientific article</a></li><li><a href="https://www.eecs.harvard.edu/~michaelm/postscripts/ReadPaper.pdf">How to read a research paper</a></li><li><a href="https://resources.nu.edu/researchprocess/readingscientificarticle">Reading a scientific article</a></li></ul></li><li> </li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice.</p><p>00:56 - Anatomy of a paper<br>02:38 - Most common advice<br>05:24 - Reading sparsity and path<br>07:30 - Notes and motivation</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7392212/">Ten simple rules for reading a scientific paper</a></li><li>Best sources I found<ul><li><a href="https://subjectguides.york.ac.uk/critical/articles">Let's get critical: Reading academic articles</a></li><li><a href="https://www.scientifica.uk.com/neurowire/gradhacks-a-guide-to-reading-research-papers">#GradHacks: A guide to reading research papers</a></li><li><a href="https://www.lib.purdue.edu/sites/default/files/libraries/engr/Tutorials/Newest%20Scientific%20Paper.pdf">How to read a scientific paper</a> (presentation)</li></ul></li><li>Some more sources<ul><li><a href="https://www.owlnet.rice.edu/~cainproj/courses/HowToReadSciArticle.pdf">How to read a scientific article</a></li><li><a href="https://www.eecs.harvard.edu/~michaelm/postscripts/ReadPaper.pdf">How to read a research paper</a></li><li><a href="https://resources.nu.edu/researchprocess/readingscientificarticle">Reading a scientific article</a></li></ul></li><li> </li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 12 Feb 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/36d075b9/8a9a30fe.mp3" length="9027946" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>565</itunes:duration>
      <itunes:summary>Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice.00:56 - Anatomy of a paper02:38 - Most common advice05:24 - Reading sparsity and path07:30 - Notes and motivationLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.Ten simple rules for reading a scientific paperBest sources I foundLet's get critical: Reading academic articles#GradHacks: A guide to reading research papersHow to read a scientific paper (presentation)Some more sourcesHow to read a scientific articleHow to read a research paperReading a scientific article</itunes:summary>
      <itunes:subtitle>Before I begin with the paper-distillation based minisodes, I figured we would go over best practices for reading research papers. I go through the anatomy of typical papers, and some generally applicable advice.00:56 - Anatomy of a paper02:38 - Most comm</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>HACKATHON: Evals November 2023 (2)</title>
      <itunes:episode>11</itunes:episode>
      <podcast:episode>11</podcast:episode>
      <itunes:title>HACKATHON: Evals November 2023 (2)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">5e003e1c-4e32-487d-905f-9672cc241a26</guid>
      <link>https://kairos.fm/intoaisafety/e011</link>
      <description>
        <![CDATA[<p>Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation.</p><p>Check out Stellaric's <a href="https://stellaric.pw/">website</a>, or follow them on <a href="https://twitter.com/stellaricpw">Twitter</a>.</p><p>01:53 - Meeting starts<br>05:05 - Pitch: extension of locked models<br>23:23 - Pitch: retroactive holdout datasets<br>34:04 - Preliminary results<br>37:44 - Next steps<br>42:55 - Recap</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://github.com/LRudL/evalugator">Evalugator</a> library</li><li><a href="https://www.alignmentforum.org/posts/rZs6ddqNnW8LXuJqA/password-locked-models-a-stress-case-for-capabilities">Password Locked Model</a> blogpost</li><li><a href="https://arxiv.org/abs/2109.07958">TruthfulQA: Measuring How Models Mimic Human Falsehoods</a></li><li><a href="https://aclanthology.org/P02-1040.pdf">BLEU: a Method for Automatic Evaluation of Machine Translation</a></li><li><a href="https://arxiv.org/abs/1905.10044">BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions</a></li><li><a href="https://arxiv.org/abs/2310.16789">Detecting Pretraining Data from Large Language Models</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation.</p><p>Check out Stellaric's <a href="https://stellaric.pw/">website</a>, or follow them on <a href="https://twitter.com/stellaricpw">Twitter</a>.</p><p>01:53 - Meeting starts<br>05:05 - Pitch: extension of locked models<br>23:23 - Pitch: retroactive holdout datasets<br>34:04 - Preliminary results<br>37:44 - Next steps<br>42:55 - Recap</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://github.com/LRudL/evalugator">Evalugator</a> library</li><li><a href="https://www.alignmentforum.org/posts/rZs6ddqNnW8LXuJqA/password-locked-models-a-stress-case-for-capabilities">Password Locked Model</a> blogpost</li><li><a href="https://arxiv.org/abs/2109.07958">TruthfulQA: Measuring How Models Mimic Human Falsehoods</a></li><li><a href="https://aclanthology.org/P02-1040.pdf">BLEU: a Method for Automatic Evaluation of Machine Translation</a></li><li><a href="https://arxiv.org/abs/1905.10044">BoolQ: Exploring the Surprising Difficulty of Natural Yes/No Questions</a></li><li><a href="https://arxiv.org/abs/2310.16789">Detecting Pretraining Data from Large Language Models</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 05 Feb 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/b1493509/6a5ceac9.mp3" length="46699824" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>2919</itunes:duration>
      <itunes:summary>Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation.Check out Stellaric's website, or follow them on Twitter.01:53 - Meeting starts05:05 - Pitch: extension of locked models23:23 - Pitch: retroactive holdout datasets34:04 - Preliminary results37:44 - Next steps42:55 - RecapLinks to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.Evalugator libraryPassword Locked Model blogpostTruthfulQA: Measuring How Models Mimic Human FalsehoodsBLEU: a Method for Automatic Evaluation of Machine TranslationBoolQ: Exploring the Surprising Difficulty of Natural Yes/No QuestionsDetecting Pretraining Data from Large Language Models</itunes:summary>
      <itunes:subtitle>Join our hackathon group for the second episode in the Evals November 2023 Hackathon subseries. In this episode, we solidify our goals for the hackathon after some preliminary experimentation and ideation.Check out Stellaric's website, or follow them on T</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Portfolios</title>
      <itunes:episode>10</itunes:episode>
      <podcast:episode>10</podcast:episode>
      <itunes:title>MINISODE: Portfolios</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">d42a5525-a155-4891-9cd1-98b1af663463</guid>
      <link>https://kairos.fm/intoaisafety/e010</link>
      <description>
        <![CDATA[<p>I provide my thoughts and recommendations regarding personal professional portfolios.</p><p>00:35 - Intro to portfolios<br>01:42 - Modern portfolios<br>02:27 - What to include<br>04:38 - Importance of visual<br>05:50 - The "About" page<br>06:25 - Tools<br>08:12 - Future of "Minisodes"</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://jime.open.ac.uk/articles/10.5334/jime.574">From Portafoglio to Eportfolio: The Evolution of Portfolio in Higher Education</a></li><li><a href="https://www.gimp.org">GIMP</a></li><li><a href="https://alternativeto.net">AlternativeTo</a></li><li><a href="https://jekyllrb.com">Jekyll</a></li><li><a href="https://pages.github.com">GitHub Pages</a></li><li><a href="https://mmistakes.github.io/minimal-mistakes/">Minimal Mistakes</a></li><li><a href="https://jacob-haimes.github.io">My portfolio</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>I provide my thoughts and recommendations regarding personal professional portfolios.</p><p>00:35 - Intro to portfolios<br>01:42 - Modern portfolios<br>02:27 - What to include<br>04:38 - Importance of visual<br>05:50 - The "About" page<br>06:25 - Tools<br>08:12 - Future of "Minisodes"</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://jime.open.ac.uk/articles/10.5334/jime.574">From Portafoglio to Eportfolio: The Evolution of Portfolio in Higher Education</a></li><li><a href="https://www.gimp.org">GIMP</a></li><li><a href="https://alternativeto.net">AlternativeTo</a></li><li><a href="https://jekyllrb.com">Jekyll</a></li><li><a href="https://pages.github.com">GitHub Pages</a></li><li><a href="https://mmistakes.github.io/minimal-mistakes/">Minimal Mistakes</a></li><li><a href="https://jacob-haimes.github.io">My portfolio</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 29 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/0258007f/e8a8e309.mp3" length="9251905" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>579</itunes:duration>
      <itunes:summary>I provide my thoughts and recommendations regarding personal professional portfolios.
00:35 - Intro to portfolios01:42 - Modern portfolios02:27 - What to include04:38 - Importance of visual05:50 - The "About" page06:25 - Tools08:12 - Future of "Minisodes"
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.From Portafoglio to Eportfolio: The Evolution of Portfolio in Higher EducationGIMPAlternativeToJekyllGitHub PagesMinimal MistakesMy portfolio</itunes:summary>
      <itunes:subtitle>I provide my thoughts and recommendations regarding personal professional portfolios.
00:35 - Intro to portfolios01:42 - Modern portfolios02:27 - What to include04:38 - Importance of visual05:50 - The "About" page06:25 - Tools08:12 - Future of "Minisodes</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>INTERVIEW: Polysemanticity w/ Dr. Darryl Wright</title>
      <itunes:episode>9</itunes:episode>
      <podcast:episode>9</podcast:episode>
      <itunes:title>INTERVIEW: Polysemanticity w/ Dr. Darryl Wright</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">f4ca1b13-2c6d-4c18-a458-9131448382b3</guid>
      <link>https://kairos.fm/intoaisafety/e009</link>
      <description>
        <![CDATA[<p>Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks.</p><p>Check out a <a href="https://into-ai-safety.github.io/episode/interview/episode-8/">diagram</a> of the decoder task used for our research!</p><p>01:46 - Interview begins<br>02:14 - Supernovae classification<br>08:58 - Penalizing polysemanticity<br>20:58 - Our "toy model"<br>30:06 - Task description<br>32:47 - Addressing hurdles<br>39:20 - Lessons learned</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.zooniverse.org">Zooniverse</a></li><li><a href="https://bluedot.org">BlueDot Impact</a></li><li><a href="https://www.aisafetysupport.org">AI Safety Support</a></li><li><a href="https://distill.pub/2020/circuits/zoom-in/">Zoom In: An Introduction to Circuits</a></li><li><a href="https://paperswithcode.com/dataset/mnist">MNIST</a> dataset on PapersWithCode</li><li><a href="https://arxiv.org/abs/2103.03386">Clusterability in Neural Networks</a></li><li><a href="https://www.cs.toronto.edu/~kriz/cifar.html">CIFAR-10</a> dataset</li><li><a href="https://www.effectivealtruism.org/ea-global">Effective Altruism Global</a></li><li><a href="https://openai.com/research/clip">CLIP</a> (blog post)</li><li><a href="https://funds.effectivealtruism.org/funds/far-future">Long Term Future Fund</a></li><li><a href="https://arxiv.org/abs/2211.09169">Engineering Monosemanticity in Toy Models</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks.</p><p>Check out a <a href="https://into-ai-safety.github.io/episode/interview/episode-8/">diagram</a> of the decoder task used for our research!</p><p>01:46 - Interview begins<br>02:14 - Supernovae classification<br>08:58 - Penalizing polysemanticity<br>20:58 - Our "toy model"<br>30:06 - Task description<br>32:47 - Addressing hurdles<br>39:20 - Lessons learned</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.zooniverse.org">Zooniverse</a></li><li><a href="https://bluedot.org">BlueDot Impact</a></li><li><a href="https://www.aisafetysupport.org">AI Safety Support</a></li><li><a href="https://distill.pub/2020/circuits/zoom-in/">Zoom In: An Introduction to Circuits</a></li><li><a href="https://paperswithcode.com/dataset/mnist">MNIST</a> dataset on PapersWithCode</li><li><a href="https://arxiv.org/abs/2103.03386">Clusterability in Neural Networks</a></li><li><a href="https://www.cs.toronto.edu/~kriz/cifar.html">CIFAR-10</a> dataset</li><li><a href="https://www.effectivealtruism.org/ea-global">Effective Altruism Global</a></li><li><a href="https://openai.com/research/clip">CLIP</a> (blog post)</li><li><a href="https://funds.effectivealtruism.org/funds/far-future">Long Term Future Fund</a></li><li><a href="https://arxiv.org/abs/2211.09169">Engineering Monosemanticity in Toy Models</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 22 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/0afbbf3c/a966e060.mp3" length="43331933" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>2709</itunes:duration>
      <itunes:summary>Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks.
Check out a diagram of the decoder task used for our research!
01:46 - Interview begins02:14 - Supernovae classification08:58 - Penalizing polysemanticity20:58 - Our "toy model"30:06 - Task description32:47 - Addressing hurdles39:20 - Lessons learned
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.

 Zooniverse
 BlueDot Impact
 AI Safety Support
 Zoom In: An Introduction to Circuits
  MNIST dataset on PapersWithCode
  Clusterability in Neural Networks
  CIFAR-10 dataset
  Effective Altruism Global
  CLIP (blog post)
  Long Term Future Fund
  Engineering Monosemanticity in Toy Models</itunes:summary>
      <itunes:subtitle>Darryl and I discuss his background, how he became interested in machine learning, and a project we are currently working on investigating the penalization of polysemanticity during the training of neural networks.
Check out a diagram of the decoder task</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Starting a Podcast</title>
      <itunes:episode>8</itunes:episode>
      <podcast:episode>8</podcast:episode>
      <itunes:title>MINISODE: Starting a Podcast</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">24d72baa-2a3e-4ae1-8d64-2af3b7cb34bb</guid>
      <link>https://kairos.fm/intoaisafety/e008</link>
      <description>
        <![CDATA[<p>A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://podcasters.spotify.com/">Spotify for Podcasters</a></li><li><a href="https://into-ai-safety.github.io">Into AI Safety podcast website</a></li><li><a href="https://www.effectivealtruism.org/ea-global">Effective Altruism Global</a></li><li><a href="https://obsproject.com">Open Broadcaster Software (OBS)</a></li><li><a href="https://craig.chat">Craig</a></li><li><a href="https://riverside.fm">Riverside</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://podcasters.spotify.com/">Spotify for Podcasters</a></li><li><a href="https://into-ai-safety.github.io">Into AI Safety podcast website</a></li><li><a href="https://www.effectivealtruism.org/ea-global">Effective Altruism Global</a></li><li><a href="https://obsproject.com">Open Broadcaster Software (OBS)</a></li><li><a href="https://craig.chat">Craig</a></li><li><a href="https://riverside.fm">Riverside</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 15 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/4823b96d/02a05da4.mp3" length="10148026" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>635</itunes:duration>
      <itunes:summary>A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar.Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.LessWrongSpotify for PodcastersInto AI Safety podcast websiteEffective Altruism GlobalOpen Broadcaster Software (OBS)CraigRiverside</itunes:summary>
      <itunes:subtitle>A summary and reflections on the path I have taken to get this podcast started, including some resources recommendations for others who want to do something similar.Links to all articles/papers which are mentioned throughout the episode can be found below</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>HACKATHON: Evals November 2023 (1)</title>
      <itunes:episode>7</itunes:episode>
      <podcast:episode>7</podcast:episode>
      <itunes:title>HACKATHON: Evals November 2023 (1)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">cbde84b2-3937-4512-8d71-fee6c23e2ec6</guid>
      <link>https://kairos.fm/intoaisafety/e007</link>
      <description>
        <![CDATA[<p>This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of the day, turned out to be pretty good.</p><p><br></p><p>Check out <a href="https://apartresearch.com">Apart Research</a>, the group that runs the <a href="https://alignmentjam.com">AlignmentJamz Hackathons</a>.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://arxiv.org/abs/2311.07723">Generalization Analogies: A Testbed for Generalizing AI Oversight to Hard-To-Measure Domains</a><ul><li><a href="https://www.lesswrong.com/posts/Yio4nmD8JMttx9o9S/new-paper-shows-truthfulness-and-instruction-following-don-t">New paper shows truthfulness &amp; instruction-following don't generalize by default</a></li><li><a href="https://joshuaclymer.github.io/generalization-analogies-website/">Generalization Analogies Website</a></li></ul></li><li><a href="https://arxiv.org/abs/2212.09251">Discovering Language Model Behaviors with Model-Written Evaluations</a><ul><li><a href="https://www.evals.anthropic.com">Model-Written Evals Website</a></li></ul></li><li><a href="https://github.com/openai/evals/tree/main/evals/elsuite/make_me_say">OpenAI Evals GitHub</a></li><li><a href="https://metr.org%20">METR</a> (previously ARC Evals)</li><li><a href="https://en.wikipedia.org/wiki/Goodhart%27s_law">Goodharting on Wikipedia</a></li><li><a href="https://arxiv.org/abs/2308.12014">From Instructions to Intrinsic Human Values, a Survey of Alignment Goals for Big Models</a></li><li><a href="https://arxiv.org/abs/2310.03693">Fine Tuning Aligned Language Models Compromises Safety Even When Users Do Not Intend</a></li><li><a href="https://arxiv.org/abs/2310.02949">Shadow Alignment: The Ease of Subverting Safely Aligned Language Models</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will Releasing the Weights of Future Large Language Models Grant Widespread Access to Pandemic Agents?</a></li><li><a href="https://www.sciencedirect.com/science/article/pii/S2666389923002210">Building Less Flawed Metrics, Understanding and Creating Better Measurement and Incentive Systems</a></li><li><a href="https://github.com/EleutherAI/lm-evaluation-harness">eLeutherAI's Model Evaluation Harness</a></li><li><a href="https://github.com/danbraunai/evalugator/tree/main">Evalugator Library</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of the day, turned out to be pretty good.</p><p><br></p><p>Check out <a href="https://apartresearch.com">Apart Research</a>, the group that runs the <a href="https://alignmentjam.com">AlignmentJamz Hackathons</a>.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li><a href="https://arxiv.org/abs/2311.07723">Generalization Analogies: A Testbed for Generalizing AI Oversight to Hard-To-Measure Domains</a><ul><li><a href="https://www.lesswrong.com/posts/Yio4nmD8JMttx9o9S/new-paper-shows-truthfulness-and-instruction-following-don-t">New paper shows truthfulness &amp; instruction-following don't generalize by default</a></li><li><a href="https://joshuaclymer.github.io/generalization-analogies-website/">Generalization Analogies Website</a></li></ul></li><li><a href="https://arxiv.org/abs/2212.09251">Discovering Language Model Behaviors with Model-Written Evaluations</a><ul><li><a href="https://www.evals.anthropic.com">Model-Written Evals Website</a></li></ul></li><li><a href="https://github.com/openai/evals/tree/main/evals/elsuite/make_me_say">OpenAI Evals GitHub</a></li><li><a href="https://metr.org%20">METR</a> (previously ARC Evals)</li><li><a href="https://en.wikipedia.org/wiki/Goodhart%27s_law">Goodharting on Wikipedia</a></li><li><a href="https://arxiv.org/abs/2308.12014">From Instructions to Intrinsic Human Values, a Survey of Alignment Goals for Big Models</a></li><li><a href="https://arxiv.org/abs/2310.03693">Fine Tuning Aligned Language Models Compromises Safety Even When Users Do Not Intend</a></li><li><a href="https://arxiv.org/abs/2310.02949">Shadow Alignment: The Ease of Subverting Safely Aligned Language Models</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will Releasing the Weights of Future Large Language Models Grant Widespread Access to Pandemic Agents?</a></li><li><a href="https://www.sciencedirect.com/science/article/pii/S2666389923002210">Building Less Flawed Metrics, Understanding and Creating Better Measurement and Incentive Systems</a></li><li><a href="https://github.com/EleutherAI/lm-evaluation-harness">eLeutherAI's Model Evaluation Harness</a></li><li><a href="https://github.com/danbraunai/evalugator/tree/main">Evalugator Library</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 08 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/04e9b1a9/e76d06ac.mp3" length="65892090" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>4119</itunes:duration>
      <itunes:summary>This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of the day, turned out to be pretty good.Check out Apart Research, the group that runs the AlignmentJamz Hackathons.Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.Generalization Analogies: A Testbed for Generalizing AI Oversight to Hard-To-Measure DomainsNew paper shows truthfulness &amp;amp; instruction-following don't generalize by defaultGeneralization Analogies WebsiteDiscovering Language Model Behaviors with Model-Written EvaluationsModel-Written Evals WebsiteOpenAI Evals GitHubMETR (previously ARC Evals)Goodharting on WikipediaFrom Instructions to Intrinsic Human Values, a Survey of Alignment Goals for Big ModelsFine Tuning Aligned Language Models Compromises Safety Even When Users Do Not IntendShadow Alignment: The Ease of Subverting Safely Aligned Language ModelsWill Releasing the Weights of Future Large Language Models Grant Widespread Access to Pandemic Agents?Building Less Flawed Metrics, Understanding and Creating Better Measurement and Incentive SystemseLeutherAI's Model Evaluation HarnessEvalugator Library</itunes:summary>
      <itunes:subtitle>This episode kicks off our first subseries, which will consist of recordings taken during my team's meetings for the AlignmentJams Evals Hackathon in November of 2023. Our team won first place, so you'll be listening to the process which, at the end of th</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Staying Up-to-Date in AI</title>
      <itunes:episode>6</itunes:episode>
      <podcast:episode>6</podcast:episode>
      <itunes:title>MINISODE: Staying Up-to-Date in AI</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">affbff67-b344-480f-bdf4-d62c1ee3e82c</guid>
      <link>https://kairos.fm/intoaisafety/e006</link>
      <description>
        <![CDATA[<p>In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the future.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li>Tools<ul><li><a href="https://feedly.com">Feedly</a></li><li><a href="https://arxiv-sanity-lite.com">arXiv Sanity Lite</a></li><li><a href="https://www.zotero.org">Zotero</a></li><li><a href="https://alternativeto.net">AlternativeTo</a></li></ul></li><li>My "Distilled AI" Folder<ul><li><a href="https://www.youtube.com/channel/UCNJ1Ymd5yFuUPtn21xtRbbw">AI Explained</a> YouTube channel</li><li><a href="https://newsletter.safe.ai">AI Safety newsletter</a></li><li><a href="https://datamachina.com">Data Machina</a> newsletter</li><li><a href="https://importai.substack.com">Import AI</a></li><li><a href="https://midwitalignment.substack.com">Midwit Alignment</a></li></ul></li><li>Honourable Mentions<ul><li><a href="https://www.alignmentforum.org">AI Alignment Forum</a></li><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://bounded-regret.ghost.io">Bounded Regret</a> (Jacob Steinhart's blog)</li><li><a href="https://www.cold-takes.com">Cold Takes</a> (Holden Karnofsky's blog)</li><li><a href="https://colah.github.io">Chris Olah's blog</a></li><li><a href="https://timdettmers.com">Tim Dettmers blog</a></li><li><a href="https://epochai.org/blog">Epoch blog</a></li><li><a href="https://www.apolloresearch.ai/blog">Apollo Research blog</a></li></ul></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the future.</p><p><br></p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.</p><ul><li>Tools<ul><li><a href="https://feedly.com">Feedly</a></li><li><a href="https://arxiv-sanity-lite.com">arXiv Sanity Lite</a></li><li><a href="https://www.zotero.org">Zotero</a></li><li><a href="https://alternativeto.net">AlternativeTo</a></li></ul></li><li>My "Distilled AI" Folder<ul><li><a href="https://www.youtube.com/channel/UCNJ1Ymd5yFuUPtn21xtRbbw">AI Explained</a> YouTube channel</li><li><a href="https://newsletter.safe.ai">AI Safety newsletter</a></li><li><a href="https://datamachina.com">Data Machina</a> newsletter</li><li><a href="https://importai.substack.com">Import AI</a></li><li><a href="https://midwitalignment.substack.com">Midwit Alignment</a></li></ul></li><li>Honourable Mentions<ul><li><a href="https://www.alignmentforum.org">AI Alignment Forum</a></li><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://bounded-regret.ghost.io">Bounded Regret</a> (Jacob Steinhart's blog)</li><li><a href="https://www.cold-takes.com">Cold Takes</a> (Holden Karnofsky's blog)</li><li><a href="https://colah.github.io">Chris Olah's blog</a></li><li><a href="https://timdettmers.com">Tim Dettmers blog</a></li><li><a href="https://epochai.org/blog">Epoch blog</a></li><li><a href="https://www.apolloresearch.ai/blog">Apollo Research blog</a></li></ul></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 01 Jan 2024 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/8dc3abe2/7c94b204.mp3" length="12554492" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>785</itunes:duration>
      <itunes:summary>In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the future.

Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance.
ToolsFeedlyarXiv Sanity LiteZoteroAlternativeToMy "Distilled AI" FolderAI Explained YouTube channelAI Safety newsletterData Machina newsletterImport AIMidwit AlignmentHonourable MentionsAI Alignment ForumLessWrongBounded Regret (Jacob Steinhart's blog)Cold Takes (Holden Karnofsky's blog)Chris Olah's blog Tim Dettmers blogEpoch blogApollo Research blog</itunes:summary>
      <itunes:subtitle>In this minisode I give some tips for staying up-to-date in the everchanging landscape of AI. I would like to point out that I am constantly iterating on these strategies, tools, and sources, so it is likely that I will make an update episode in the futur</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>INTERVIEW: Applications w/ Alice Rigg</title>
      <itunes:episode>5</itunes:episode>
      <podcast:episode>5</podcast:episode>
      <itunes:title>INTERVIEW: Applications w/ Alice Rigg</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">b2215bea-18f8-497a-b76f-4bb095c79cf0</guid>
      <link>https://kairos.fm/intoaisafety/e005</link>
      <description>
        <![CDATA[<p>Alice Rigg, a mechanistic interpretability researcher from Ottawa, Canada, joins me to discuss their path and the applications process for research/mentorship programs.</p><p>Join the <a href="https://discord.gg/KNxJwaytTG">Mech Interp Discord server</a> and attend reading groups at 11:00am on Wednesdays (Mountain Time)!</p><p>Check out Alice's <a href="https://woog97.github.io">website</a>.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://www.eleuther.ai">EleutherAI</a><br> <ul><li>Join the public <a href="https://www.discord.gg/eleutherai">EleutherAI discord server</a></li></ul></li><li> </li><li><a href="https://distill.pub">Distill</a></li><li><a href="https://www.effectivealtruism.org">Effective Altruism</a> (EA)</li><li><a href="https://www.lesswrong.com/posts/zwf68YaySvXhWYCdh/mats-summer-2023-retrospective">MATS Retrospective Summer 2023</a> post</li><li><a href="https://docs.google.com/document/d/1jce3f64Fz7PXmdCEyd9i0PTmcFaiP1pZdcBn5ye5sxY/edit">Ambitious Mechanistic Interpretability</a> AISC research plan by Alice Rigg</li><li><a href="https://berkeleyaisafety.com/spar">SPAR</a></li><li><a href="https://stability.ai">Stability AI</a><br> <ul><li>During their most recent fundraising round, Stability AI had a valuation of $4B <a href="https://www.bloomberg.com/news/articles/2023-03-03/openai-rival-stable-diffusion-maker-seeks-to-raise-funds-at-4-billion-valuation">(Bloomberg)</a></li></ul></li><li> </li><li><a href="https://discord.gg/KNxJwaytTG">Mech Interp Discord Server</a></li></ul><p><br></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Alice Rigg, a mechanistic interpretability researcher from Ottawa, Canada, joins me to discuss their path and the applications process for research/mentorship programs.</p><p>Join the <a href="https://discord.gg/KNxJwaytTG">Mech Interp Discord server</a> and attend reading groups at 11:00am on Wednesdays (Mountain Time)!</p><p>Check out Alice's <a href="https://woog97.github.io">website</a>.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://www.eleuther.ai">EleutherAI</a><br> <ul><li>Join the public <a href="https://www.discord.gg/eleutherai">EleutherAI discord server</a></li></ul></li><li> </li><li><a href="https://distill.pub">Distill</a></li><li><a href="https://www.effectivealtruism.org">Effective Altruism</a> (EA)</li><li><a href="https://www.lesswrong.com/posts/zwf68YaySvXhWYCdh/mats-summer-2023-retrospective">MATS Retrospective Summer 2023</a> post</li><li><a href="https://docs.google.com/document/d/1jce3f64Fz7PXmdCEyd9i0PTmcFaiP1pZdcBn5ye5sxY/edit">Ambitious Mechanistic Interpretability</a> AISC research plan by Alice Rigg</li><li><a href="https://berkeleyaisafety.com/spar">SPAR</a></li><li><a href="https://stability.ai">Stability AI</a><br> <ul><li>During their most recent fundraising round, Stability AI had a valuation of $4B <a href="https://www.bloomberg.com/news/articles/2023-03-03/openai-rival-stable-diffusion-maker-seeks-to-raise-funds-at-4-billion-valuation">(Bloomberg)</a></li></ul></li><li> </li><li><a href="https://discord.gg/KNxJwaytTG">Mech Interp Discord Server</a></li></ul><p><br></p>]]>
      </content:encoded>
      <pubDate>Mon, 18 Dec 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/2889532d/02620130.mp3" length="67843965" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>4241</itunes:duration>
      <itunes:summary>Alice Rigg, a mechanistic interpretability researcher from Ottawa, Canada, joins me to discuss their path and the applications process for research/mentorship programs.
Join the Mech Interp Discord server and attend reading groups at 11:00am on Wednesdays (Mountain Time)!
Check out Alice's website.
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. 

 EleutherAI
 Join the public EleutherAI discord server

Distill
 Effective Altruism (EA)
 MATS Retrospective Summer 2023 post
 Ambitious Mechanistic Interpretability AISC research plan by Alice Rigg
 SPAR
  Stability AI
  During their most recent fundraising round, Stability AI had a valuation of $4B (Bloomberg)

Mech Interp Discord Server</itunes:summary>
      <itunes:subtitle>Alice Rigg, a mechanistic interpretability researcher from Ottawa, Canada, joins me to discuss their path and the applications process for research/mentorship programs.
Join the Mech Interp Discord server and attend reading groups at 11:00am on Wednesday</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: Program Applications (Winter 2024)</title>
      <itunes:episode>4</itunes:episode>
      <podcast:episode>4</podcast:episode>
      <itunes:title>MINISODE: Program Applications (Winter 2024)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">2f5c7fd0-e665-4850-a48c-b4d20f05f4af</guid>
      <link>https://kairos.fm/intoaisafety/e004</link>
      <description>
        <![CDATA[<p>We're back after a month-long hiatus with a podcast refactor and advice on the applications process for research/mentorship programs.</p><p>Check out the <a href="https://into-ai-safety.github.io/about/">About</a> page on the <a href="https://into-ai-safety.github.io">Into AI Safety</a> website for a summary of the logistics updates.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://www.matsprogram.org">MATS</a></li><li><a href="https://www.constellation.org/programs/astra-fellowship">ASTRA Fellowship</a></li><li><a href="https://www.arena.education">ARENA</a></li><li><a href="https://aisafety.camp">AI Safety Camp</a></li><li><a href="https://www.bluedotimpact.org">BlueDot Impact</a></li><li><a href="https://www.techwithtim.net">Tech with Tim</a></li><li>Fast.AI's <a href="https://course.fast.ai">Practical Deep Learning for Coders</a></li><li><a href="https://www.kaggle.com">Kaggle</a></li><li><a href="https://alignmentjam.com">AlignmentJams</a></li><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://www.alignmentforum.org">AI Alignment Forum</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>We're back after a month-long hiatus with a podcast refactor and advice on the applications process for research/mentorship programs.</p><p>Check out the <a href="https://into-ai-safety.github.io/about/">About</a> page on the <a href="https://into-ai-safety.github.io">Into AI Safety</a> website for a summary of the logistics updates.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://www.matsprogram.org">MATS</a></li><li><a href="https://www.constellation.org/programs/astra-fellowship">ASTRA Fellowship</a></li><li><a href="https://www.arena.education">ARENA</a></li><li><a href="https://aisafety.camp">AI Safety Camp</a></li><li><a href="https://www.bluedotimpact.org">BlueDot Impact</a></li><li><a href="https://www.techwithtim.net">Tech with Tim</a></li><li>Fast.AI's <a href="https://course.fast.ai">Practical Deep Learning for Coders</a></li><li><a href="https://www.kaggle.com">Kaggle</a></li><li><a href="https://alignmentjam.com">AlignmentJams</a></li><li><a href="https://www.lesswrong.com">LessWrong</a></li><li><a href="https://www.alignmentforum.org">AI Alignment Forum</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 11 Dec 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/6a54c881/0ef512b8.mp3" length="17356182" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>1085</itunes:duration>
      <itunes:summary>We're back after a month-long hiatus with a podcast refactor and advice on the applications process for research/mentorship programs.
Check out the About page on the Into AI Safety website for a summary of the logistics updates.
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. MATSASTRA FellowshipARENAAI Safety Camp
 BlueDot ImpactTech with TimFast.AI's Practical Deep Learning for CodersKaggleAlignmentJamsLessWrongAI Alignment Forum</itunes:summary>
      <itunes:subtitle>We're back after a month-long hiatus with a podcast refactor and advice on the applications process for research/mentorship programs.
Check out the About page on the Into AI Safety website for a summary of the logistics updates.
Links to all articles/pa</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
      <podcast:person role="Editor" href="https://intoaisafety.transistor.fm/people/chase-precopia">Chase Precopia</podcast:person>
    </item>
    <item>
      <title>MINISODE: EAG Takeaways (Boston 2023)</title>
      <itunes:episode>3</itunes:episode>
      <podcast:episode>3</podcast:episode>
      <itunes:title>MINISODE: EAG Takeaways (Boston 2023)</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">dc2fc55a-dca7-44c6-bf9e-be0c4e9eb971</guid>
      <link>https://kairos.fm/intoaisafety/e003</link>
      <description>
        <![CDATA[<p>This episode is a brief overview of the major takeaways I had from attending EAG Boston 2023, and an update on my plans for the podcast moving forward.</p><p><strong>TL;DL</strong></p><ul><li>Starting in early December (2023), I will be uploading episodes on a biweekly basis (day TBD).</li><li>I won't be releasing another episode until then, so that I can build a cache of episodes up.</li><li>During this month (November 2023), I'll also try to get the podcast up on more platforms, set up comments on more platforms, and create an anonymous feedback form.</li></ul><p><strong>Links</strong> Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://forum.effectivealtruism.org/posts/8R2NffQiCsn3F7hpv/how-to-generate-research-proposals">How to generate research proposals</a></li><li><a href="https://www.effectivealtruism.org/articles/karolina-sarek-how-to-do-research-that-matters">Karolina Sarek: How to do research that matters</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will releasing the weights of future large language models grant widespread access to pandemic agents?</a></li></ul><p>Like the show? Think it could be improved? Fill out <a href="https://docs.google.com/forms/d/e/1FAIpQLSdXqpGGb0uWgpQd8CUvKP6g2Ki8FrDsEBzFfQBrBoSZPlxjDQ/viewform?usp=sf_link">this anonymous feedback form </a>to let me know!</p><p>Please email all inquiries to intoaisafety@gmail.com.</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>This episode is a brief overview of the major takeaways I had from attending EAG Boston 2023, and an update on my plans for the podcast moving forward.</p><p><strong>TL;DL</strong></p><ul><li>Starting in early December (2023), I will be uploading episodes on a biweekly basis (day TBD).</li><li>I won't be releasing another episode until then, so that I can build a cache of episodes up.</li><li>During this month (November 2023), I'll also try to get the podcast up on more platforms, set up comments on more platforms, and create an anonymous feedback form.</li></ul><p><strong>Links</strong> Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://forum.effectivealtruism.org/posts/8R2NffQiCsn3F7hpv/how-to-generate-research-proposals">How to generate research proposals</a></li><li><a href="https://www.effectivealtruism.org/articles/karolina-sarek-how-to-do-research-that-matters">Karolina Sarek: How to do research that matters</a></li><li><a href="https://arxiv.org/abs/2310.18233">Will releasing the weights of future large language models grant widespread access to pandemic agents?</a></li></ul><p>Like the show? Think it could be improved? Fill out <a href="https://docs.google.com/forms/d/e/1FAIpQLSdXqpGGb0uWgpQd8CUvKP6g2Ki8FrDsEBzFfQBrBoSZPlxjDQ/viewform?usp=sf_link">this anonymous feedback form </a>to let me know!</p><p>Please email all inquiries to intoaisafety@gmail.com.</p>]]>
      </content:encoded>
      <pubDate>Mon, 04 Dec 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/101ea918/e54e49cc.mp3" length="9228951" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>577</itunes:duration>
      <itunes:summary>This episode is a brief overview of the major takeaways I had from attending EAG Boston 2023, and an update on my plans for the podcast moving forward.
TL;DL Starting in early December (2023), I will be uploading episodes on a biweekly basis (day TBD).I won't be releasing another episode until then, so that I can build a cache of episodes up.During this month (November 2023), I'll also try to get the podcast up on more platforms, set up comments on more platforms, and create an anonymous feedback form.
Links
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. How to generate research proposalsKarolina Sarek: How to do research that mattersWill releasing the weights of future large language models grant widespread access to pandemic agents?

Like the show? Think it could be improved? Fill out this anonymous feedback form to let me know!
Please email all inquiries to intoaisafety@gmail.com.</itunes:summary>
      <itunes:subtitle>This episode is a brief overview of the major takeaways I had from attending EAG Boston 2023, and an update on my plans for the podcast moving forward.
TL;DL Starting in early December (2023), I will be uploading episodes on a biweekly basis (day TBD).I </itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
    </item>
    <item>
      <title>FEEDBACK: AISC Proposal w/ Remmelt Ellen</title>
      <itunes:episode>2</itunes:episode>
      <podcast:episode>2</podcast:episode>
      <itunes:title>FEEDBACK: AISC Proposal w/ Remmelt Ellen</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">084356f7-066c-49d9-b8d9-a45f7162f138</guid>
      <link>https://kairos.fm/intoaisafety/e002</link>
      <description>
        <![CDATA[<p>In this episode I discuss my initial research proposal for the 2024 Winter AI Safety Camp with one of the individuals who helps facilitate the program, Remmelt Ellen.</p><p>The proposal is titled The Effect of Machine Learning on Bioengineered Pandemic Risk. A doc-capsule of the proposal at the time of this recording can be found <a href="https://docs.google.com/document/d/1bbFDNc_hzhzYqN6pn1jYvPhSRp2CyYez4UP0ly_mMRM/edit?usp=sharing">at this link</a>.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://chemrxiv.org/engage/chemrxiv/article-details/61551803d1fc335b7cf8fd45">MegaSyn: Integrating Generative Molecule Design, Automated Analog Designer and Synthetic Viability Prediction</a></li><li><a href="https://www.nature.com/articles/s42256-022-00465-9?fbclid=IwAR11_V1cd9SUxEvUfwrWMA7TUcroyYIY1nBDUL3KaS-8B4rG5MIqZCmjm0M">Dual use of artificial-intelligence-powered drug discovery</a></li><li><a href="https://arxiv.org/abs/2306.13952">Artificial intelligence and biological misuse: Differentiating risks of language models and biological design tools</a></li><li><a href="https://www.alignmentforum.org/posts/ChDH335ckdvpxXaXX/model-organisms-of-misalignment-the-case-for-a-new-pillar-of-1">Model Organisms of Misalignment: The Case for a New Pillar of Alignment Research</a></li><li><a href="https://arxiv.org/abs/2310.02949">Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models</a></li><li><a href="https://arxiv.org/abs/2310.03693">Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!</a></li><li><a href="https://www.alignmentforum.org/posts/3eqHYxfWb5x4Qfz8C/unrlhf-efficiently-undoing-llm-safeguards">unRLHF - Efficiently undoing LLM safeguards</a></li></ul>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode I discuss my initial research proposal for the 2024 Winter AI Safety Camp with one of the individuals who helps facilitate the program, Remmelt Ellen.</p><p>The proposal is titled The Effect of Machine Learning on Bioengineered Pandemic Risk. A doc-capsule of the proposal at the time of this recording can be found <a href="https://docs.google.com/document/d/1bbFDNc_hzhzYqN6pn1jYvPhSRp2CyYez4UP0ly_mMRM/edit?usp=sharing">at this link</a>.</p><p>Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. </p><ul><li><a href="https://chemrxiv.org/engage/chemrxiv/article-details/61551803d1fc335b7cf8fd45">MegaSyn: Integrating Generative Molecule Design, Automated Analog Designer and Synthetic Viability Prediction</a></li><li><a href="https://www.nature.com/articles/s42256-022-00465-9?fbclid=IwAR11_V1cd9SUxEvUfwrWMA7TUcroyYIY1nBDUL3KaS-8B4rG5MIqZCmjm0M">Dual use of artificial-intelligence-powered drug discovery</a></li><li><a href="https://arxiv.org/abs/2306.13952">Artificial intelligence and biological misuse: Differentiating risks of language models and biological design tools</a></li><li><a href="https://www.alignmentforum.org/posts/ChDH335ckdvpxXaXX/model-organisms-of-misalignment-the-case-for-a-new-pillar-of-1">Model Organisms of Misalignment: The Case for a New Pillar of Alignment Research</a></li><li><a href="https://arxiv.org/abs/2310.02949">Shadow Alignment: The Ease of Subverting Safely-Aligned Language Models</a></li><li><a href="https://arxiv.org/abs/2310.03693">Fine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!</a></li><li><a href="https://www.alignmentforum.org/posts/3eqHYxfWb5x4Qfz8C/unrlhf-efficiently-undoing-llm-safeguards">unRLHF - Efficiently undoing LLM safeguards</a></li></ul>]]>
      </content:encoded>
      <pubDate>Mon, 27 Nov 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/57154a54/62466372.mp3" length="54368967" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>3398</itunes:duration>
      <itunes:summary>In this episode I discuss my initial research proposal for the 2024 Winter AI Safety Camp with one of the individuals who helps facilitate the program, Remmelt Ellen.
The proposal is titled The Effect of Machine Learning on Bioengineered Pandemic Risk. A doc-capsule of the proposal at the time of this recording can be found at this link.
Links to all articles/papers which are mentioned throughout the episode can be found below, in order of their appearance. MegaSyn: Integrating Generative Molecule Design, Automated Analog Designer and Synthetic Viability PredictionDual use of artificial-intelligence-powered drug discoveryArtificial intelligence and biological misuse: Differentiating risks of language models and biological design toolsModel Organisms of Misalignment: The Case for a New Pillar of Alignment ResearchShadow Alignment: The Ease of Subverting Safely-Aligned Language ModelsFine-tuning Aligned Language Models Compromises Safety, Even When Users Do Not Intend To!unRLHF - Efficiently undoing LLM safeguards</itunes:summary>
      <itunes:subtitle>In this episode I discuss my initial research proposal for the 2024 Winter AI Safety Camp with one of the individuals who helps facilitate the program, Remmelt Ellen.
The proposal is titled The Effect of Machine Learning on Bioengineered Pandemic Risk. A</itunes:subtitle>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
    </item>
    <item>
      <title>MINISODE: Introduction and Motivation</title>
      <itunes:episode>1</itunes:episode>
      <podcast:episode>1</podcast:episode>
      <itunes:title>MINISODE: Introduction and Motivation</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">4a709b26-c999-4113-b23f-daeac781b2df</guid>
      <link>https://kairos.fm/intoaisafety/e001</link>
      <description>
        <![CDATA[<p>Welcome to the <strong>Into AI Safety</strong> podcast! In this episode I provide reasoning for why I am starting this podcast, what I am trying to accomplish with it, and a little bit of background on how I got here.</p><p>Please email all inquiries and suggestions to intoaisafety@gmail.com.</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Welcome to the <strong>Into AI Safety</strong> podcast! In this episode I provide reasoning for why I am starting this podcast, what I am trying to accomplish with it, and a little bit of background on how I got here.</p><p>Please email all inquiries and suggestions to intoaisafety@gmail.com.</p>]]>
      </content:encoded>
      <pubDate>Mon, 13 Nov 2023 09:00:00 -0700</pubDate>
      <author>Jacob Haimes</author>
      <enclosure url="https://op3.dev/e/media.transistor.fm/2b9dbb68/7bb95cf1.mp3" length="9326333" type="audio/mpeg"/>
      <itunes:author>Jacob Haimes</itunes:author>
      <itunes:duration>583</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>Welcome to the <strong>Into AI Safety</strong> podcast! In this episode I provide reasoning for why I am starting this podcast, what I am trying to accomplish with it, and a little bit of background on how I got here.</p><p>Please email all inquiries and suggestions to intoaisafety@gmail.com.</p>]]>
      </itunes:summary>
      <itunes:keywords>technology, AI, artificial intelligence, machine learning, safety</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="https://jacob-haimes.github.io" img="https://img.transistor.fm/6X2qc0h5t6L29JagHPBUMkc0dLKnOVeeQpxv1FIfF4s/rs:fill:800:800:1/q:60/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS8xZmEw/MTM2M2RhZjQwNDRm/M2MxZTY3MWIyZDNj/MmYyOS5wbmc.jpg">Jacob Haimes</podcast:person>
    </item>
  </channel>
</rss>
