<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet href="/stylesheet.xsl" type="text/xsl"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:podcast="https://podcastindex.org/namespace/1.0">
  <channel>
    <atom:link rel="self" type="application/atom+xml" href="https://feeds.transistor.fm/tic-tac-toe-the-hard-way" title="MP3 Audio"/>
    <atom:link rel="hub" href="https://pubsubhubbub.appspot.com/"/>
    <podcast:podping usesPodping="true"/>
    <title>Tic-Tac-Toe the Hard Way </title>
    <generator>Transistor (https://transistor.fm)</generator>
    <itunes:new-feed-url>https://feeds.transistor.fm/tic-tac-toe-the-hard-way</itunes:new-feed-url>
    <description>A writer and a software engineer from Google's People + AI Research team explore the human choices that shape machine learning systems by building competing tic-tac-toe agents.</description>
    <copyright>© 2025 People + AI Research</copyright>
    <podcast:guid>48949ff6-4f44-5b9f-a5c5-68d75bd22aa1</podcast:guid>
    <podcast:locked owner="thehardway@google.com">no</podcast:locked>
    <podcast:trailer pubdate="Tue, 21 Jul 2020 16:57:00 -0400" url="https://media.transistor.fm/1da39dd3/60c20d04.mp3" length="5237428" type="audio/mpeg">Introducing Tic-Tac-Toe the Hard Way </podcast:trailer>
    <language>en</language>
    <pubDate>Fri, 09 Feb 2024 15:09:10 -0500</pubDate>
    <lastBuildDate>Tue, 02 Dec 2025 16:28:18 -0500</lastBuildDate>
    <link>http://pair.withgoogle.com/thehardway/</link>
    <image>
      <url>https://img.transistor.fm/QPOQXD2Zv1iUyQ69H3aaIJ_o9QQ4mW3-RszZfOWpbxk/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9zaG93/LzEyNDcxLzE1OTQ4/NDU3MDctYXJ0d29y/ay5qcGc.jpg</url>
      <title>Tic-Tac-Toe the Hard Way </title>
      <link>http://pair.withgoogle.com/thehardway/</link>
    </image>
    <itunes:category text="Technology"/>
    <itunes:category text="Education"/>
    <itunes:type>serial</itunes:type>
    <itunes:author>People + AI Research</itunes:author>
    <itunes:image href="https://img.transistor.fm/QPOQXD2Zv1iUyQ69H3aaIJ_o9QQ4mW3-RszZfOWpbxk/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9zaG93/LzEyNDcxLzE1OTQ4/NDU3MDctYXJ0d29y/ay5qcGc.jpg"/>
    <itunes:summary>A writer and a software engineer from Google's People + AI Research team explore the human choices that shape machine learning systems by building competing tic-tac-toe agents.</itunes:summary>
    <itunes:subtitle>A writer and a software engineer from Google's People + AI Research team explore the human choices that shape machine learning systems by building competing tic-tac-toe agents..</itunes:subtitle>
    <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
    <itunes:owner>
      <itunes:name>Lucas Dixon</itunes:name>
    </itunes:owner>
    <itunes:complete>No</itunes:complete>
    <itunes:explicit>No</itunes:explicit>
    <item>
      <title>Introducing Tic-Tac-Toe the Hard Way </title>
      <itunes:title>Introducing Tic-Tac-Toe the Hard Way </itunes:title>
      <itunes:episodeType>trailer</itunes:episodeType>
      <guid isPermaLink="false">4caa7322-6e50-45af-856d-5535c0e7973f</guid>
      <link>https://share.transistor.fm/s/1da39dd3</link>
      <description>
        <![CDATA[<p>Introducing the podcast where a writer and a software engineer explore the human choices that shape machine learning systems by building competing tic-tac-toe agents. Brought to you by Google's <a href="https://pair.withgoogle.com/">People + AI Research</a> team.</p><p>More at: <a href="http://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway</a></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Introducing the podcast where a writer and a software engineer explore the human choices that shape machine learning systems by building competing tic-tac-toe agents. Brought to you by Google's <a href="https://pair.withgoogle.com/">People + AI Research</a> team.</p><p>More at: <a href="http://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway</a></p>]]>
      </content:encoded>
      <pubDate>Tue, 21 Jul 2020 16:57:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/1da39dd3/60c20d04.mp3" length="5237428" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>129</itunes:duration>
      <itunes:summary>Introducing the podcast where a writer and a software engineer explore the human choices that shape machine learning systems by building competing tic-tac-toe agents. Brought to you by Google's People + AI Research team.</itunes:summary>
      <itunes:subtitle>Introducing the podcast where a writer and a software engineer explore the human choices that shape machine learning systems by building competing tic-tac-toe agents. Brought to you by Google's People + AI Research team.</itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>Howdy, and the myth of “pouring in data”</title>
      <itunes:episode>1</itunes:episode>
      <podcast:episode>1</podcast:episode>
      <itunes:title>Howdy, and the myth of “pouring in data”</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">ff26c8b4-ce61-4f92-90cd-cbc6793f4326</guid>
      <link>https://share.transistor.fm/s/6768a641</link>
      <description>
        <![CDATA[<p>Welcome to the podcast! We’re Yannick and David, a software engineer and a non-technical writer. Over the next 9 episodes we’re going to use two different approaches to build machine learning systems that play two versions of tic-tac-toe. Building a machine learning app requires humans making a lot of decisions. We start by agreeing that David will use a “supervised learning” approach while Yannick will go with “reinforcement learning.”</p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway</a>. </p><p><br></p><p>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Welcome to the podcast! We’re Yannick and David, a software engineer and a non-technical writer. Over the next 9 episodes we’re going to use two different approaches to build machine learning systems that play two versions of tic-tac-toe. Building a machine learning app requires humans making a lot of decisions. We start by agreeing that David will use a “supervised learning” approach while Yannick will go with “reinforcement learning.”</p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway</a>. </p><p><br></p><p>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </content:encoded>
      <pubDate>Tue, 21 Jul 2020 17:05:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/6768a641/3c90deac.mp3" length="52953191" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1321</itunes:duration>
      <itunes:summary>David and Yannick get started on their project to build competing machine learning systems that play tic-tac-toe. They discuss the human choices that will shape their systems along the way. </itunes:summary>
      <itunes:subtitle>David and Yannick get started on their project to build competing machine learning systems that play tic-tac-toe. They discuss the human choices that will shape their systems along the way. </itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>What does a tic-tac-toe board look like to machine learning?</title>
      <itunes:episode>2</itunes:episode>
      <podcast:episode>2</podcast:episode>
      <itunes:title>What does a tic-tac-toe board look like to machine learning?</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">e6e611a1-a894-4dd5-a0f1-99cfbab37428</guid>
      <link>https://share.transistor.fm/s/eba6db79</link>
      <description>
        <![CDATA[<p>How should David represent the data needed to train his machine learning system? What does a tic-tac-toe board “look” like to ML? Should he train it on games or on individual boards? How does this decision affect how and how well the machine will learn to play? Plus, an intro to reinforcement learning, the approach Yannick will be taking.</p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>How should David represent the data needed to train his machine learning system? What does a tic-tac-toe board “look” like to ML? Should he train it on games or on individual boards? How does this decision affect how and how well the machine will learn to play? Plus, an intro to reinforcement learning, the approach Yannick will be taking.</p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </content:encoded>
      <pubDate>Wed, 22 Jul 2020 09:38:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/eba6db79/215fdfcf.mp3" length="56336654" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1406</itunes:duration>
      <itunes:summary>David delves into questions around data and training for his model including: What does a tic-tac-toe board “look” like to ML? Plus, an intro to reinforcement learning, the approach Yannick will be taking.</itunes:summary>
      <itunes:subtitle>David delves into questions around data and training for his model including: What does a tic-tac-toe board “look” like to ML? Plus, an intro to reinforcement learning, the approach Yannick will be taking.</itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>From tic-tac-toe moves to ML model</title>
      <itunes:episode>3</itunes:episode>
      <podcast:episode>3</podcast:episode>
      <itunes:title>From tic-tac-toe moves to ML model</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">305a734a-c1a4-4603-aaad-575ed51ea202</guid>
      <link>https://share.transistor.fm/s/7b0d33a4</link>
      <description>
        <![CDATA[<p>Once we have the data we need—thousands of sample games--how do we turn it into something the ML can train itself on? That means understanding how training works, and what a model is. </p><p><strong>Resources:</strong><br><a href="https://developers.google.com/machine-learning/glossary#one-hot_encoding">See a definition of one-hot encoding</a></p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Once we have the data we need—thousands of sample games--how do we turn it into something the ML can train itself on? That means understanding how training works, and what a model is. </p><p><strong>Resources:</strong><br><a href="https://developers.google.com/machine-learning/glossary#one-hot_encoding">See a definition of one-hot encoding</a></p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </content:encoded>
      <pubDate>Wed, 22 Jul 2020 09:55:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/7b0d33a4/9fb903d4.mp3" length="51985890" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1297</itunes:duration>
      <itunes:summary>Once we have the data we need—thousands of sample games—how do we turn it into something the ML can train itself on? That means understanding how training works, and what a model is. </itunes:summary>
      <itunes:subtitle>Once we have the data we need—thousands of sample games—how do we turn it into something the ML can train itself on? That means understanding how training works, and what a model is. </itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>Beating random: What it means to have trained a model</title>
      <itunes:episode>4</itunes:episode>
      <podcast:episode>4</podcast:episode>
      <itunes:title>Beating random: What it means to have trained a model</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">fdccbad0-3827-4ef6-af2a-3d527c7ea96f</guid>
      <link>https://share.transistor.fm/s/65ec8e69</link>
      <description>
        <![CDATA[<p>David did it! He trained a machine learning model to play tic-tac-toe! (Well, with lots of help from Yannick.) How did the whole training experience go? How do you tell how training went? How did his model do against a player that makes random tic-tac-toe moves? </p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>David did it! He trained a machine learning model to play tic-tac-toe! (Well, with lots of help from Yannick.) How did the whole training experience go? How do you tell how training went? How did his model do against a player that makes random tic-tac-toe moves? </p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </content:encoded>
      <pubDate>Wed, 22 Jul 2020 11:30:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/65ec8e69/cce2db19.mp3" length="41446936" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1034</itunes:duration>
      <itunes:summary>David did it! He trained a machine learning model to play tic-tac-toe! How did his model do against a player that makes random tic-tac-toe moves? </itunes:summary>
      <itunes:subtitle>David did it! He trained a machine learning model to play tic-tac-toe! How did his model do against a player that makes random tic-tac-toe moves? </itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>Give that model a treat! : Reinforcement learning explained</title>
      <itunes:episode>5</itunes:episode>
      <podcast:episode>5</podcast:episode>
      <itunes:title>Give that model a treat! : Reinforcement learning explained</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">aca34995-31df-4a19-8f4a-a43e0a5c3e5b</guid>
      <link>https://share.transistor.fm/s/730e69f9</link>
      <description>
        <![CDATA[<p>Switching gears, we focus on how Yannick’s been training his model using reinforcement learning.  He explains the differences from David’s supervised learning approach. We find out how his system performs against a player that makes random tic-tac-toe moves.</p><p><strong>Resources: </strong></p><p><a href="https://www.manning.com/books/deep-learning-with-javascript">Deep Learning for JavaScript book</a></p><p><a href="https://arxiv.org/abs/1312.5602">Playing Atari with Deep Reinforcement Learning</a></p><p><a href="https://www.youtube.com/watch?v=V1eYniJ0Rnk&amp;vl=en">Two Minute Papers episode on Atari DQN</a></p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br></p><p>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p><p><br></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Switching gears, we focus on how Yannick’s been training his model using reinforcement learning.  He explains the differences from David’s supervised learning approach. We find out how his system performs against a player that makes random tic-tac-toe moves.</p><p><strong>Resources: </strong></p><p><a href="https://www.manning.com/books/deep-learning-with-javascript">Deep Learning for JavaScript book</a></p><p><a href="https://arxiv.org/abs/1312.5602">Playing Atari with Deep Reinforcement Learning</a></p><p><a href="https://www.youtube.com/watch?v=V1eYniJ0Rnk&amp;vl=en">Two Minute Papers episode on Atari DQN</a></p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br></p><p>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p><p><br></p>]]>
      </content:encoded>
      <pubDate>Wed, 22 Jul 2020 11:35:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/730e69f9/3384a6ed.mp3" length="62637240" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1564</itunes:duration>
      <itunes:summary>Switching gears, we focus on how Yannick’s been training his model using reinforcement learning.  He explains the differences from David’s supervised learning approach. We find out how his system performs against a player that makes random tic-tac-toe moves.</itunes:summary>
      <itunes:subtitle>Switching gears, we focus on how Yannick’s been training his model using reinforcement learning.  He explains the differences from David’s supervised learning approach. We find out how his system performs against a player that makes random tic-tac-toe mov</itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>Head to Head: the Big ML Smackdown!</title>
      <itunes:episode>6</itunes:episode>
      <podcast:episode>6</podcast:episode>
      <itunes:title>Head to Head: the Big ML Smackdown!</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">1521ddac-190c-43f6-949c-017e131b1587</guid>
      <link>https://share.transistor.fm/s/e2baaaf8</link>
      <description>
        <![CDATA[<p>David and Yannick’s tic-tac-toe ML agents face-off against each other in tic-tac-toe!</p><p><a href="https://pair.withgoogle.com/thehardway/tic-tac-toe/viewer/"><strong>See the agents play each other</strong></a><strong>!</strong></p><p><br></p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br></p><p>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p><p><br></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>David and Yannick’s tic-tac-toe ML agents face-off against each other in tic-tac-toe!</p><p><a href="https://pair.withgoogle.com/thehardway/tic-tac-toe/viewer/"><strong>See the agents play each other</strong></a><strong>!</strong></p><p><br></p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br></p><p>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p><p><br></p>]]>
      </content:encoded>
      <pubDate>Wed, 22 Jul 2020 11:45:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/e2baaaf8/8a7376f2.mp3" length="60851294" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1519</itunes:duration>
      <itunes:summary>David and Yannick’s tic-tac-toe ML agents face-off against each other in tic-tac-toe!</itunes:summary>
      <itunes:subtitle>David and Yannick’s tic-tac-toe ML agents face-off against each other in tic-tac-toe!</itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>Enter tic-tac-two</title>
      <itunes:episode>7</itunes:episode>
      <podcast:episode>7</podcast:episode>
      <itunes:title>Enter tic-tac-two</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">5195b4a2-20d1-4a46-a93d-4add4f9ba6b4</guid>
      <link>https://share.transistor.fm/s/f0d5fead</link>
      <description>
        <![CDATA[<p>David’s variant of tic-tac-toe that we’re calling tic-tac-two is only slightly different but turns out to be far more complex. This requires rethinking what the ML system will need in order to learn how to play, and  how to represent that data.</p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>David’s variant of tic-tac-toe that we’re calling tic-tac-two is only slightly different but turns out to be far more complex. This requires rethinking what the ML system will need in order to learn how to play, and  how to represent that data.</p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </content:encoded>
      <pubDate>Wed, 22 Jul 2020 11:50:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/f0d5fead/d1d7cb97.mp3" length="51284252" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1280</itunes:duration>
      <itunes:summary>David’s variant of tic-tac-toe that we’re calling tic-tac-two is only slightly different but turns out to be far more complex. This requires rethinking what the ML system will need in order to learn how to play, and  how to represent that data.</itunes:summary>
      <itunes:subtitle>David’s variant of tic-tac-toe that we’re calling tic-tac-two is only slightly different but turns out to be far more complex. This requires rethinking what the ML system will need in order to learn how to play, and  how to represent that data.</itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>Head to Head: The Even Bigger ML Smackdown!</title>
      <itunes:episode>8</itunes:episode>
      <podcast:episode>8</podcast:episode>
      <itunes:title>Head to Head: The Even Bigger ML Smackdown!</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">5c97fe83-09c8-4b10-899d-fee0f8efb60f</guid>
      <link>https://share.transistor.fm/s/32397484</link>
      <description>
        <![CDATA[<p>Yannick and David’s systems play against each other in 500 games. Who’s going to win? And what can we learn about how the ML may be working by thinking about the results?</p><p><a href="https://pair.withgoogle.com/thehardway/tic-tac-two/viewer/">See the agents play each other in Tic-Tac-Two</a>!</p><p><br></p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Yannick and David’s systems play against each other in 500 games. Who’s going to win? And what can we learn about how the ML may be working by thinking about the results?</p><p><a href="https://pair.withgoogle.com/thehardway/tic-tac-two/viewer/">See the agents play each other in Tic-Tac-Two</a>!</p><p><br></p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </content:encoded>
      <pubDate>Wed, 22 Jul 2020 11:55:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/32397484/9524ea88.mp3" length="58725064" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1466</itunes:duration>
      <itunes:summary>Yannick and David’s systems play against each other in 500 games. Who’s going to win? And what can we learn about how the ML may be working by thinking about the results?</itunes:summary>
      <itunes:subtitle>Yannick and David’s systems play against each other in 500 games. Who’s going to win? And what can we learn about how the ML may be working by thinking about the results?</itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
    <item>
      <title>Lessons learned</title>
      <itunes:episode>9</itunes:episode>
      <podcast:episode>9</podcast:episode>
      <itunes:title>Lessons learned</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">f016b63e-a017-421d-ba19-6eb72a96c81b</guid>
      <link>https://share.transistor.fm/s/2f59a2be</link>
      <description>
        <![CDATA[<p>What have we learned about machine learning and the human decisions that shape it? And is machine learning perhaps changing our minds about how the world outside of machine learning — also known as the world — works?</p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>What have we learned about machine learning and the human decisions that shape it? And is machine learning perhaps changing our minds about how the world outside of machine learning — also known as the world — works?</p><p>For more information about the show, check out <a href="https://pair.withgoogle.com/thehardway/">pair.withgoogle.com/thehardway/</a>.</p><p><br>You can reach out to the hosts on Twitter: <a href="https://twitter.com/dweinberger">@dweinberger</a> and <a href="https://twitter.com/tafsiri">@tafsiri</a>. </p>]]>
      </content:encoded>
      <pubDate>Wed, 22 Jul 2020 11:58:00 -0400</pubDate>
      <author>People + AI Research</author>
      <enclosure url="https://media.transistor.fm/2f59a2be/be359f28.mp3" length="79315236" type="audio/mpeg"/>
      <itunes:author>People + AI Research</itunes:author>
      <itunes:duration>1981</itunes:duration>
      <itunes:summary>What have we learned about machine learning and the human decisions that shape it? And is machine learning perhaps changing our minds about how the world outside of machine learning — also known as the world — works?</itunes:summary>
      <itunes:subtitle>What have we learned about machine learning and the human decisions that shape it? And is machine learning perhaps changing our minds about how the world outside of machine learning — also known as the world — works?</itunes:subtitle>
      <itunes:keywords>machine learning, ML, AI, human-centered, reinforcement learning, supervised learning, tic-tac-toe, games, Google, tic tac toe</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
    </item>
  </channel>
</rss>
