<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet href="/stylesheet.xsl" type="text/xsl"?>
<rss version="2.0" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:sy="http://purl.org/rss/1.0/modules/syndication/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:podcast="https://podcastindex.org/namespace/1.0">
  <channel>
    <atom:link rel="self" type="application/atom+xml" href="https://feeds.transistor.fm/computer-vision-decoded" title="MP3 Audio"/>
    <atom:link rel="hub" href="https://pubsubhubbub.appspot.com/"/>
    <podcast:podping usesPodping="true"/>
    <title>Computer Vision Decoded</title>
    <generator>Transistor (https://transistor.fm)</generator>
    <itunes:new-feed-url>https://feeds.transistor.fm/computer-vision-decoded</itunes:new-feed-url>
    <description>A tidal wave of computer vision innovation is quickly having an impact on everyone's lives, but not everyone has the time to sit down and read through a bunch of news articles and learn what it means for them. In Computer Vision Decoded, we sit down with Jared Heinly, the Chief Scientist at EveryPoint, to discuss topics in today’s quickly evolving world of computer vision and decode what they mean for you. If you want to be sure you understand everything happening in the world of computer vision, don't miss an episode!</description>
    <copyright>© EveryPoint, Inc</copyright>
    <podcast:guid>1879aefd-2d8d-5196-bf1d-9ddb3f1ba4dd</podcast:guid>
    <podcast:locked owner="jonathan.stephens@everypoint.io">no</podcast:locked>
    <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
    <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    <language>en</language>
    <pubDate>Mon, 14 Jul 2025 13:59:13 -0700</pubDate>
    <lastBuildDate>Tue, 02 Dec 2025 16:22:46 -0800</lastBuildDate>
    <link>http://www.everypoint.io</link>
    <image>
      <url>https://img.transistor.fm/hOZksfhwtNK5id-6RkatzRCTlcpJzZcyTEt7KwMLUc4/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9zaG93/LzMxOTQ2LzE2NTU4/MzEwMTItYXJ0d29y/ay5qcGc.jpg</url>
      <title>Computer Vision Decoded</title>
      <link>http://www.everypoint.io</link>
    </image>
    <itunes:category text="Technology"/>
    <itunes:category text="Science"/>
    <itunes:type>episodic</itunes:type>
    <itunes:author>EveryPoint</itunes:author>
    <itunes:image href="https://img.transistor.fm/hOZksfhwtNK5id-6RkatzRCTlcpJzZcyTEt7KwMLUc4/rs:fill:0:0:1/w:1400/h:1400/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9zaG93/LzMxOTQ2LzE2NTU4/MzEwMTItYXJ0d29y/ay5qcGc.jpg"/>
    <itunes:summary>A tidal wave of computer vision innovation is quickly having an impact on everyone's lives, but not everyone has the time to sit down and read through a bunch of news articles and learn what it means for them. In Computer Vision Decoded, we sit down with Jared Heinly, the Chief Scientist at EveryPoint, to discuss topics in today’s quickly evolving world of computer vision and decode what they mean for you. If you want to be sure you understand everything happening in the world of computer vision, don't miss an episode!</itunes:summary>
    <itunes:subtitle>A tidal wave of computer vision innovation is quickly having an impact on everyone's lives, but not everyone has the time to sit down and read through a bunch of news articles and learn what it means for them.</itunes:subtitle>
    <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
    <itunes:owner>
      <itunes:name>Jonathan Stephens</itunes:name>
    </itunes:owner>
    <itunes:complete>No</itunes:complete>
    <itunes:explicit>No</itunes:explicit>
    <item>
      <title>The Evolution of Image Based 3D Reconstruction</title>
      <itunes:episode>18</itunes:episode>
      <podcast:episode>18</podcast:episode>
      <itunes:title>The Evolution of Image Based 3D Reconstruction</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">ddf07911-b9c4-48f7-9471-6a17eb0cc99f</guid>
      <link>https://share.transistor.fm/s/25600296</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we bring to you a live recording of Jared Heinly presentation on the evolution of image based 3D reconstruction. This recording was from a Computer Vision Decoded meetup in Pittsburgh with a visual component. If you would like to tune in with the visuals, we recommend watching the episode on our YouTube channel: https://youtu.be/Gwib5IcTKHI</p><p>Follow:</p><ul><li>Jared on X: https://x.com/JaredHeinly</li><li>Jonathan on X: https://x.com/jonstephens85</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we bring to you a live recording of Jared Heinly presentation on the evolution of image based 3D reconstruction. This recording was from a Computer Vision Decoded meetup in Pittsburgh with a visual component. If you would like to tune in with the visuals, we recommend watching the episode on our YouTube channel: https://youtu.be/Gwib5IcTKHI</p><p>Follow:</p><ul><li>Jared on X: https://x.com/JaredHeinly</li><li>Jonathan on X: https://x.com/jonstephens85</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Mon, 14 Jul 2025 13:59:08 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/25600296/7ecc9ca9.mp3" length="79009908" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>1975</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we bring to you a live recording of Jared Heinly presentation on the evolution of image based 3D reconstruction. This recording was from a Computer Vision Decoded meetup in Pittsburgh with a visual component. If you would like to tune in with the visuals, we recommend watching the episode on our YouTube channel: https://youtu.be/Gwib5IcTKHI</p><p>Follow:</p><ul><li>Jared on X: https://x.com/JaredHeinly</li><li>Jonathan on X: https://x.com/jonstephens85</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at https://www.everypoint.io</p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>Understanding Gaussian Splatting w/ NVIDIA's Ruilong Li</title>
      <itunes:episode>17</itunes:episode>
      <podcast:episode>17</podcast:episode>
      <itunes:title>Understanding Gaussian Splatting w/ NVIDIA's Ruilong Li</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">56374e11-3402-460c-b407-2db0a0026c3f</guid>
      <link>https://share.transistor.fm/s/48d798f2</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, our hosts Jonathan Stephens and Jared Heinly are joined by Ruilong Li, a researcher at NVIDIA and key contributor to both Nerfstudio and gsplat, to dive deep into 3D Gaussian Splatting. They explore how this relatively new technology works, from the fundamentals of gaussian representations to the optimization process that creates photorealistic 3D scenes. Ruilong explains the technical details behind gaussian splatting, and discusses the development of the popular gsplat library. The conversation covers practical advice for capturing high-quality data, the iterative training process, and how Gaussian splatting compares to other 3D representations like meshes and NeRFs.</p><p>Links: </p><ul><li>gsplat: https://github.com/nerfstudio-project/gsplat</li><li>Nerfstudio: https://docs.nerf.studio/</li></ul><p><br>Follow:</p><ul><li>Ruilong on X: https://x.com/ruilong_li</li><li>Jared on X: https://x.com/JaredHeinly</li><li>Jonathan on X: https://x.com/jonstephens85</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, our hosts Jonathan Stephens and Jared Heinly are joined by Ruilong Li, a researcher at NVIDIA and key contributor to both Nerfstudio and gsplat, to dive deep into 3D Gaussian Splatting. They explore how this relatively new technology works, from the fundamentals of gaussian representations to the optimization process that creates photorealistic 3D scenes. Ruilong explains the technical details behind gaussian splatting, and discusses the development of the popular gsplat library. The conversation covers practical advice for capturing high-quality data, the iterative training process, and how Gaussian splatting compares to other 3D representations like meshes and NeRFs.</p><p>Links: </p><ul><li>gsplat: https://github.com/nerfstudio-project/gsplat</li><li>Nerfstudio: https://docs.nerf.studio/</li></ul><p><br>Follow:</p><ul><li>Ruilong on X: https://x.com/ruilong_li</li><li>Jared on X: https://x.com/JaredHeinly</li><li>Jonathan on X: https://x.com/jonstephens85</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Tue, 17 Jun 2025 10:42:31 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/48d798f2/176afcbf.mp3" length="189334077" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>4733</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded, our hosts Jonathan Stephens and Jared Heinly are joined by Ruilong Li, a researcher at NVIDIA and key contributor to both Nerfstudio and gsplat, to dive deep into 3D Gaussian Splatting. They explore how this relatively new technology works, from the fundamentals of gaussian representations to the optimization process that creates photorealistic 3D scenes. Ruilong explains the technical details behind gaussian splatting, and discusses the development of the popular gsplat library. The conversation covers practical advice for capturing high-quality data, the iterative training process, and how Gaussian splatting compares to other 3D representations like meshes and NeRFs.</p><p>Links: </p><ul><li>gsplat: https://github.com/nerfstudio-project/gsplat</li><li>Nerfstudio: https://docs.nerf.studio/</li></ul><p><br>Follow:</p><ul><li>Ruilong on X: https://x.com/ruilong_li</li><li>Jared on X: https://x.com/JaredHeinly</li><li>Jonathan on X: https://x.com/jonstephens85</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at https://www.everypoint.io</p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, 3d, gaussian splatting, gsplat, nerfstudio, radiance fields</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:person role="Guest" href="https://www.liruilong.cn/" img="https://img.transistor.fm/lwrEg41r3bExtKwIsp9SrIchEXvlT3iwi6x-GERFcwk/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS80YjYy/YmFiNTkxMzExZDcx/NTE4YmJmOGVjZWQ4/YjM2My5wbmc.jpg">Ruilong Li</podcast:person>
    </item>
    <item>
      <title>Camera Types for 3D Reconstruction Explained</title>
      <itunes:episode>16</itunes:episode>
      <podcast:episode>16</podcast:episode>
      <itunes:title>Camera Types for 3D Reconstruction Explained</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">243e4410-e41b-4543-a60a-73e660cabee4</guid>
      <link>https://share.transistor.fm/s/172053fc</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, hosts Jonathan Stephens and Jared Heinly explore the various types of cameras used in computer vision and 3D reconstruction. They discuss the strengths and weaknesses of smartphone cameras, DSLR and mirrorless cameras, action cameras, drones, and specialized cameras like 360, thermal, and event cameras. The conversation emphasizes the importance of understanding camera specifications, metadata, and the impact of different lenses on image quality. The hosts also provide practical advice for beginners in 3D reconstruction, encouraging them to start with the cameras they already own.</p><p>Takeaways</p><ul><li>Smartphones are versatile and user-friendly for photography.</li><li>RAW images preserve more data than JPEGs, aiding in post-processing.</li><li>Mirrorless and DSLR cameras offer better low-light performance and lens flexibility.</li><li>Drones provide unique perspectives and programmable flight paths for capturing images.</li><li>360 cameras allow for quick scene capture but may require additional processing for 3D reconstruction.</li><li>Event cameras capture rapid changes in intensity, useful for robotics applications.</li><li>Thermal and multispectral cameras are specialized for specific applications, not typically used for 3D reconstruction.</li><li>Understanding camera metadata is crucial for effective image processing.</li><li>Choosing the right camera depends on the specific needs of the project.</li><li>Starting with a smartphone is a low barrier to entry for beginners in 3D reconstruction.</li></ul><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, hosts Jonathan Stephens and Jared Heinly explore the various types of cameras used in computer vision and 3D reconstruction. They discuss the strengths and weaknesses of smartphone cameras, DSLR and mirrorless cameras, action cameras, drones, and specialized cameras like 360, thermal, and event cameras. The conversation emphasizes the importance of understanding camera specifications, metadata, and the impact of different lenses on image quality. The hosts also provide practical advice for beginners in 3D reconstruction, encouraging them to start with the cameras they already own.</p><p>Takeaways</p><ul><li>Smartphones are versatile and user-friendly for photography.</li><li>RAW images preserve more data than JPEGs, aiding in post-processing.</li><li>Mirrorless and DSLR cameras offer better low-light performance and lens flexibility.</li><li>Drones provide unique perspectives and programmable flight paths for capturing images.</li><li>360 cameras allow for quick scene capture but may require additional processing for 3D reconstruction.</li><li>Event cameras capture rapid changes in intensity, useful for robotics applications.</li><li>Thermal and multispectral cameras are specialized for specific applications, not typically used for 3D reconstruction.</li><li>Understanding camera metadata is crucial for effective image processing.</li><li>Choosing the right camera depends on the specific needs of the project.</li><li>Starting with a smartphone is a low barrier to entry for beginners in 3D reconstruction.</li></ul><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Wed, 14 May 2025 08:32:22 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/172053fc/1333d555.mp3" length="72585577" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>4535</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded, hosts Jonathan Stephens and Jared Heinly explore the various types of cameras used in computer vision and 3D reconstruction. They discuss the strengths and weaknesses of smartphone cameras, DSLR and mirrorless cameras, action cameras, drones, and specialized cameras like 360, thermal, and event cameras. The conversation emphasizes the importance of understanding camera specifications, metadata, and the impact of different lenses on image quality. The hosts also provide practical advice for beginners in 3D reconstruction, encouraging them to start with the cameras they already own.</p><p>Takeaways</p><ul><li>Smartphones are versatile and user-friendly for photography.</li><li>RAW images preserve more data than JPEGs, aiding in post-processing.</li><li>Mirrorless and DSLR cameras offer better low-light performance and lens flexibility.</li><li>Drones provide unique perspectives and programmable flight paths for capturing images.</li><li>360 cameras allow for quick scene capture but may require additional processing for 3D reconstruction.</li><li>Event cameras capture rapid changes in intensity, useful for robotics applications.</li><li>Thermal and multispectral cameras are specialized for specific applications, not typically used for 3D reconstruction.</li><li>Understanding camera metadata is crucial for effective image processing.</li><li>Choosing the right camera depends on the specific needs of the project.</li><li>Starting with a smartphone is a low barrier to entry for beginners in 3D reconstruction.</li></ul><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/172053fc/transcript.txt" type="text/plain"/>
    </item>
    <item>
      <title>Understanding 3D Reconstruction with COLMAP</title>
      <itunes:episode>15</itunes:episode>
      <podcast:episode>15</podcast:episode>
      <itunes:title>Understanding 3D Reconstruction with COLMAP</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">24d59d08-3e94-4a39-9bd6-ca4ed0886ed6</guid>
      <link>https://share.transistor.fm/s/606d8d4b</link>
      <description>
        <![CDATA[<p>In this episode, Jonathan Stephens and Jared Heinly delve into the intricacies of COLMAP, a powerful tool for 3D reconstruction from images. They discuss the workflow of COLMAP, including feature extraction, correspondence search, incremental reconstruction, and the importance of camera models. The conversation also covers advanced topics like geometric verification, bundle adjustment, and the newer GLOMAP method, which offers a faster alternative to traditional reconstruction techniques. Listeners are encouraged to experiment with COLMAP and learn through hands-on experience.</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode, Jonathan Stephens and Jared Heinly delve into the intricacies of COLMAP, a powerful tool for 3D reconstruction from images. They discuss the workflow of COLMAP, including feature extraction, correspondence search, incremental reconstruction, and the importance of camera models. The conversation also covers advanced topics like geometric verification, bundle adjustment, and the newer GLOMAP method, which offers a faster alternative to traditional reconstruction techniques. Listeners are encouraged to experiment with COLMAP and learn through hands-on experience.</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Thu, 03 Apr 2025 11:03:05 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/606d8d4b/b39af011.mp3" length="54777171" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3422</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode, Jonathan Stephens and Jared Heinly delve into the intricacies of COLMAP, a powerful tool for 3D reconstruction from images. They discuss the workflow of COLMAP, including feature extraction, correspondence search, incremental reconstruction, and the importance of camera models. The conversation also covers advanced topics like geometric verification, bundle adjustment, and the newer GLOMAP method, which offers a faster alternative to traditional reconstruction techniques. Listeners are encouraged to experiment with COLMAP and learn through hands-on experience.</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </itunes:summary>
      <itunes:keywords>COLMAP, 3D reconstruction, computer vision, feature extraction, incremental reconstruction, GLOMAP, camera poses, geometric verification, bundle adjustment, image matching</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/606d8d4b/transcript.srt" type="application/x-subrip" rel="captions"/>
    </item>
    <item>
      <title>Tips and Tricks for 3D Reconstruction in Different Environments</title>
      <itunes:episode>14</itunes:episode>
      <podcast:episode>14</podcast:episode>
      <itunes:title>Tips and Tricks for 3D Reconstruction in Different Environments</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">635b3bcc-181a-4c33-9660-03f26ab7367f</guid>
      <link>https://share.transistor.fm/s/484ffb14</link>
      <description>
        <![CDATA[<p>In this episode, we discuss practical tips and challenges in 3D reconstruction from images, focusing on various environments such as urban, indoor, and outdoor settings. We explore issues like repetitive structures, lighting conditions, and the impact of reflections and shadows on reconstruction quality. The conversation also touches on the importance of camera motion, lens distortion, and the role of machine learning in enhancing reconstruction processes. Listeners gain insights into optimizing their 3D capture techniques for better results.</p><p>Key Takeaways</p><ul><li>Repetitive structures can confuse computer vision algorithms.</li><li>Lighting conditions greatly affect image quality and reconstruction accuracy.</li><li>Wide-angle lenses can help capture more unique features.</li><li>Indoor environments present unique challenges like textureless walls.</li><li>Aerial imaging requires careful management of lens distortion.</li><li>Understanding the application context is crucial for effective 3D reconstruction.</li><li>Camera motion should be varied to avoid distortion and drift.</li><li>Planning captures based on goals can lead to better results.</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at <a href="https://www.youtube.com/redirect?event=video_description&amp;redir_token=QUFFLUhqa1NUdnFnSzJ2RmlENUVydnU4SEJ5Wi1KbUtLZ3xBQ3Jtc0tscFc4QjRhalp2eHlrSWYwTEQ0OTVZNnNqalNlN3RYMzZId0FMdWZldTBHY0RSTUdROVdBSHctXzQ0TVVaTFo3a0NWcFV3cUswMnMwakdjVkNPMEkxei1RM0lGUFhla1pXMkV2U3ZUQ0xHRFlWcnRoMA&amp;q=https%3A%2F%2Fwww.everypoint.io%2F&amp;v=1LbKcnlzTUU">https://www.everypoint.io</a></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode, we discuss practical tips and challenges in 3D reconstruction from images, focusing on various environments such as urban, indoor, and outdoor settings. We explore issues like repetitive structures, lighting conditions, and the impact of reflections and shadows on reconstruction quality. The conversation also touches on the importance of camera motion, lens distortion, and the role of machine learning in enhancing reconstruction processes. Listeners gain insights into optimizing their 3D capture techniques for better results.</p><p>Key Takeaways</p><ul><li>Repetitive structures can confuse computer vision algorithms.</li><li>Lighting conditions greatly affect image quality and reconstruction accuracy.</li><li>Wide-angle lenses can help capture more unique features.</li><li>Indoor environments present unique challenges like textureless walls.</li><li>Aerial imaging requires careful management of lens distortion.</li><li>Understanding the application context is crucial for effective 3D reconstruction.</li><li>Camera motion should be varied to avoid distortion and drift.</li><li>Planning captures based on goals can lead to better results.</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at <a href="https://www.youtube.com/redirect?event=video_description&amp;redir_token=QUFFLUhqa1NUdnFnSzJ2RmlENUVydnU4SEJ5Wi1KbUtLZ3xBQ3Jtc0tscFc4QjRhalp2eHlrSWYwTEQ0OTVZNnNqalNlN3RYMzZId0FMdWZldTBHY0RSTUdROVdBSHctXzQ0TVVaTFo3a0NWcFV3cUswMnMwakdjVkNPMEkxei1RM0lGUFhla1pXMkV2U3ZUQ0xHRFlWcnRoMA&amp;q=https%3A%2F%2Fwww.everypoint.io%2F&amp;v=1LbKcnlzTUU">https://www.everypoint.io</a></p>]]>
      </content:encoded>
      <pubDate>Tue, 04 Mar 2025 10:50:18 -0800</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/484ffb14/f3ea2fd4.mp3" length="78149051" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>4883</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode, we discuss practical tips and challenges in 3D reconstruction from images, focusing on various environments such as urban, indoor, and outdoor settings. We explore issues like repetitive structures, lighting conditions, and the impact of reflections and shadows on reconstruction quality. The conversation also touches on the importance of camera motion, lens distortion, and the role of machine learning in enhancing reconstruction processes. Listeners gain insights into optimizing their 3D capture techniques for better results.</p><p>Key Takeaways</p><ul><li>Repetitive structures can confuse computer vision algorithms.</li><li>Lighting conditions greatly affect image quality and reconstruction accuracy.</li><li>Wide-angle lenses can help capture more unique features.</li><li>Indoor environments present unique challenges like textureless walls.</li><li>Aerial imaging requires careful management of lens distortion.</li><li>Understanding the application context is crucial for effective 3D reconstruction.</li><li>Camera motion should be varied to avoid distortion and drift.</li><li>Planning captures based on goals can lead to better results.</li></ul><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at <a href="https://www.youtube.com/redirect?event=video_description&amp;redir_token=QUFFLUhqa1NUdnFnSzJ2RmlENUVydnU4SEJ5Wi1KbUtLZ3xBQ3Jtc0tscFc4QjRhalp2eHlrSWYwTEQ0OTVZNnNqalNlN3RYMzZId0FMdWZldTBHY0RSTUdROVdBSHctXzQ0TVVaTFo3a0NWcFV3cUswMnMwakdjVkNPMEkxei1RM0lGUFhla1pXMkV2U3ZUQ0xHRFlWcnRoMA&amp;q=https%3A%2F%2Fwww.everypoint.io%2F&amp;v=1LbKcnlzTUU">https://www.everypoint.io</a></p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>Exploring Depth Maps in Computer Vision</title>
      <itunes:episode>13</itunes:episode>
      <podcast:episode>13</podcast:episode>
      <itunes:title>Exploring Depth Maps in Computer Vision</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">97ea628c-3339-4165-a3dc-a53c242f6a96</guid>
      <link>https://share.transistor.fm/s/1f2e632c</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, Jonathan Stephens and Jared Heinly explore the concept of depth maps in computer vision. They discuss the basics of depth and depth maps, their applications in smartphones, and the various types of depth maps. The conversation delves into the role of depth maps in photogrammetry and 3D reconstruction, as well as future trends in depth sensing and machine learning. The episode highlights the importance of depth maps in enhancing photography, gaming, and autonomous systems.</p><p>Key Takeaways:</p><ul><li>Depth maps represent how far away objects are from a sensor.</li><li>Smartphones use depth maps for features like portrait mode.</li><li>There are multiple types of depth maps, including absolute and relative.</li><li>Depth maps are essential in photogrammetry for creating 3D models.</li><li>Machine learning is increasingly used for depth estimation.</li><li>Depth maps can be generated from various sensors, including LiDAR.</li><li>The resolution and baseline of cameras affect depth perception.</li><li>Depth maps are used in gaming for rendering and performance optimization.</li><li>Sensor fusion combines data from multiple sources for better accuracy.</li><li>The future of depth sensing will likely involve more machine learning applications.</li></ul><p><br>Episode Chapters<br>00:00 Introduction to Depth Maps</p><p>00:13 Understanding Depth in Computer Vision</p><p>06:52 Applications of Depth Maps in Photography</p><p>07:53 Types of Depth Maps Created by Smartphones</p><p>08:31 Depth Measurement Techniques</p><p>16:00 Machine Learning and Depth Estimation</p><p>19:18 Absolute vs Relative Depth Maps</p><p>23:14 Disparity Maps and Depth Ordering</p><p>26:53 Depth Maps in Graphics and Gaming</p><p>31:24 Depth Maps in Photogrammetry</p><p>34:12 Utilizing Depth Maps in 3D Reconstruction</p><p>37:51 Sensor Fusion and SLAM Technologies</p><p>41:31 Future Trends in Depth Sensing</p><p>46:37 Innovations in Computational Photography</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at <a href="https://www.youtube.com/redirect?event=video_description&amp;redir_token=QUFFLUhqa1NUdnFnSzJ2RmlENUVydnU4SEJ5Wi1KbUtLZ3xBQ3Jtc0tscFc4QjRhalp2eHlrSWYwTEQ0OTVZNnNqalNlN3RYMzZId0FMdWZldTBHY0RSTUdROVdBSHctXzQ0TVVaTFo3a0NWcFV3cUswMnMwakdjVkNPMEkxei1RM0lGUFhla1pXMkV2U3ZUQ0xHRFlWcnRoMA&amp;q=https%3A%2F%2Fwww.everypoint.io%2F&amp;v=1LbKcnlzTUU">https://www.everypoint.io</a></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, Jonathan Stephens and Jared Heinly explore the concept of depth maps in computer vision. They discuss the basics of depth and depth maps, their applications in smartphones, and the various types of depth maps. The conversation delves into the role of depth maps in photogrammetry and 3D reconstruction, as well as future trends in depth sensing and machine learning. The episode highlights the importance of depth maps in enhancing photography, gaming, and autonomous systems.</p><p>Key Takeaways:</p><ul><li>Depth maps represent how far away objects are from a sensor.</li><li>Smartphones use depth maps for features like portrait mode.</li><li>There are multiple types of depth maps, including absolute and relative.</li><li>Depth maps are essential in photogrammetry for creating 3D models.</li><li>Machine learning is increasingly used for depth estimation.</li><li>Depth maps can be generated from various sensors, including LiDAR.</li><li>The resolution and baseline of cameras affect depth perception.</li><li>Depth maps are used in gaming for rendering and performance optimization.</li><li>Sensor fusion combines data from multiple sources for better accuracy.</li><li>The future of depth sensing will likely involve more machine learning applications.</li></ul><p><br>Episode Chapters<br>00:00 Introduction to Depth Maps</p><p>00:13 Understanding Depth in Computer Vision</p><p>06:52 Applications of Depth Maps in Photography</p><p>07:53 Types of Depth Maps Created by Smartphones</p><p>08:31 Depth Measurement Techniques</p><p>16:00 Machine Learning and Depth Estimation</p><p>19:18 Absolute vs Relative Depth Maps</p><p>23:14 Disparity Maps and Depth Ordering</p><p>26:53 Depth Maps in Graphics and Gaming</p><p>31:24 Depth Maps in Photogrammetry</p><p>34:12 Utilizing Depth Maps in 3D Reconstruction</p><p>37:51 Sensor Fusion and SLAM Technologies</p><p>41:31 Future Trends in Depth Sensing</p><p>46:37 Innovations in Computational Photography</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at <a href="https://www.youtube.com/redirect?event=video_description&amp;redir_token=QUFFLUhqa1NUdnFnSzJ2RmlENUVydnU4SEJ5Wi1KbUtLZ3xBQ3Jtc0tscFc4QjRhalp2eHlrSWYwTEQ0OTVZNnNqalNlN3RYMzZId0FMdWZldTBHY0RSTUdROVdBSHctXzQ0TVVaTFo3a0NWcFV3cUswMnMwakdjVkNPMEkxei1RM0lGUFhla1pXMkV2U3ZUQ0xHRFlWcnRoMA&amp;q=https%3A%2F%2Fwww.everypoint.io%2F&amp;v=1LbKcnlzTUU">https://www.everypoint.io</a></p>]]>
      </content:encoded>
      <pubDate>Tue, 18 Feb 2025 11:23:40 -0800</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/1f2e632c/9280228f.mp3" length="55230653" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3451</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded, Jonathan Stephens and Jared Heinly explore the concept of depth maps in computer vision. They discuss the basics of depth and depth maps, their applications in smartphones, and the various types of depth maps. The conversation delves into the role of depth maps in photogrammetry and 3D reconstruction, as well as future trends in depth sensing and machine learning. The episode highlights the importance of depth maps in enhancing photography, gaming, and autonomous systems.</p><p>Key Takeaways:</p><ul><li>Depth maps represent how far away objects are from a sensor.</li><li>Smartphones use depth maps for features like portrait mode.</li><li>There are multiple types of depth maps, including absolute and relative.</li><li>Depth maps are essential in photogrammetry for creating 3D models.</li><li>Machine learning is increasingly used for depth estimation.</li><li>Depth maps can be generated from various sensors, including LiDAR.</li><li>The resolution and baseline of cameras affect depth perception.</li><li>Depth maps are used in gaming for rendering and performance optimization.</li><li>Sensor fusion combines data from multiple sources for better accuracy.</li><li>The future of depth sensing will likely involve more machine learning applications.</li></ul><p><br>Episode Chapters<br>00:00 Introduction to Depth Maps</p><p>00:13 Understanding Depth in Computer Vision</p><p>06:52 Applications of Depth Maps in Photography</p><p>07:53 Types of Depth Maps Created by Smartphones</p><p>08:31 Depth Measurement Techniques</p><p>16:00 Machine Learning and Depth Estimation</p><p>19:18 Absolute vs Relative Depth Maps</p><p>23:14 Disparity Maps and Depth Ordering</p><p>26:53 Depth Maps in Graphics and Gaming</p><p>31:24 Depth Maps in Photogrammetry</p><p>34:12 Utilizing Depth Maps in 3D Reconstruction</p><p>37:51 Sensor Fusion and SLAM Technologies</p><p>41:31 Future Trends in Depth Sensing</p><p>46:37 Innovations in Computational Photography</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services. Learn more at <a href="https://www.youtube.com/redirect?event=video_description&amp;redir_token=QUFFLUhqa1NUdnFnSzJ2RmlENUVydnU4SEJ5Wi1KbUtLZ3xBQ3Jtc0tscFc4QjRhalp2eHlrSWYwTEQ0OTVZNnNqalNlN3RYMzZId0FMdWZldTBHY0RSTUdROVdBSHctXzQ0TVVaTFo3a0NWcFV3cUswMnMwakdjVkNPMEkxei1RM0lGUFhla1pXMkV2U3ZUQ0xHRFlWcnRoMA&amp;q=https%3A%2F%2Fwww.everypoint.io%2F&amp;v=1LbKcnlzTUU">https://www.everypoint.io</a></p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/1f2e632c/transcript.txt" type="text/plain"/>
    </item>
    <item>
      <title>What's New in 2025 for Computer Vision?</title>
      <itunes:episode>12</itunes:episode>
      <podcast:episode>12</podcast:episode>
      <itunes:title>What's New in 2025 for Computer Vision?</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">27c660f0-6a30-4033-8e69-edb99101f9af</guid>
      <link>https://share.transistor.fm/s/daa5e3a7</link>
      <description>
        <![CDATA[<p>After an 18 month hiatus, we are back! In this episode of Computer Vision Decoded, hosts Jonathan Stephens and Jared Heinly discuss the latest advancements in computer vision technology, personal updates, and insights from the industry. They explore topics such as real-time 3D reconstruction, computer vision research, SLAM, event cameras, and the impact of generative AI on robotics. The conversation highlights the importance of merging traditional techniques with modern machine learning approaches to solve real-world problems effectively.</p><p><strong>Chapters</strong></p><p>00:00 Intro &amp; Personal Updates<br>04:36 Real-Time 3D Reconstruction on iPhones<br>09:40 Advancements in SfM<br>14:56 Event Cameras<br>17:39 Neural Networks in 3D Reconstruction<br>26:30 SLAM and Machine Learning Innovation<br>29:48 Applications of SLAM in Robotics<br>34:19 NVIDIA's Cosmos and Physical AI<br>40:18 Generative AI for Real-World Applications<br>43:50 The Future of Gaussian Splatting and 3D Reconstruction</p><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>After an 18 month hiatus, we are back! In this episode of Computer Vision Decoded, hosts Jonathan Stephens and Jared Heinly discuss the latest advancements in computer vision technology, personal updates, and insights from the industry. They explore topics such as real-time 3D reconstruction, computer vision research, SLAM, event cameras, and the impact of generative AI on robotics. The conversation highlights the importance of merging traditional techniques with modern machine learning approaches to solve real-world problems effectively.</p><p><strong>Chapters</strong></p><p>00:00 Intro &amp; Personal Updates<br>04:36 Real-Time 3D Reconstruction on iPhones<br>09:40 Advancements in SfM<br>14:56 Event Cameras<br>17:39 Neural Networks in 3D Reconstruction<br>26:30 SLAM and Machine Learning Innovation<br>29:48 Applications of SLAM in Robotics<br>34:19 NVIDIA's Cosmos and Physical AI<br>40:18 Generative AI for Real-World Applications<br>43:50 The Future of Gaussian Splatting and 3D Reconstruction</p><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Tue, 11 Feb 2025 09:30:13 -0800</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/daa5e3a7/d5fdc5a4.mp3" length="48067652" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3003</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>After an 18 month hiatus, we are back! In this episode of Computer Vision Decoded, hosts Jonathan Stephens and Jared Heinly discuss the latest advancements in computer vision technology, personal updates, and insights from the industry. They explore topics such as real-time 3D reconstruction, computer vision research, SLAM, event cameras, and the impact of generative AI on robotics. The conversation highlights the importance of merging traditional techniques with modern machine learning approaches to solve real-world problems effectively.</p><p><strong>Chapters</strong></p><p>00:00 Intro &amp; Personal Updates<br>04:36 Real-Time 3D Reconstruction on iPhones<br>09:40 Advancements in SfM<br>14:56 Event Cameras<br>17:39 Neural Networks in 3D Reconstruction<br>26:30 SLAM and Machine Learning Innovation<br>29:48 Applications of SLAM in Robotics<br>34:19 NVIDIA's Cosmos and Physical AI<br>40:18 Generative AI for Real-World Applications<br>43:50 The Future of Gaussian Splatting and 3D Reconstruction</p><p><br>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>A Computer Vision Scientist Reacts to the iPhone 15 Announcement</title>
      <itunes:episode>11</itunes:episode>
      <podcast:episode>11</podcast:episode>
      <itunes:title>A Computer Vision Scientist Reacts to the iPhone 15 Announcement</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">84df81ff-1093-40bb-97d7-8b91f5d4a2ba</guid>
      <link>https://share.transistor.fm/s/34856bf3</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into our in-house computer vision expert's reaction to the iPhone 15 and iPhone 15 Pro announcement.</p><p>We dive into the camera upgrades, decode what a quad sensor means, and even talk about the importance of depth maps.</p><p>Episode timeline:</p><p>00:00 Intro<br>02:59 iPhone 15 Overview<br>05:15 iPhone 15 Main Camera<br>07:20 Quad Pixel Sensor Explained<br>15:45 Depth Maps Explained<br>22:57 iPhone 15 Pro Overview<br>27:01 iPhone 15 Pro Cameras<br>32:20 Spatial Video<br>36:00 A17 Pro Chipset</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into our in-house computer vision expert's reaction to the iPhone 15 and iPhone 15 Pro announcement.</p><p>We dive into the camera upgrades, decode what a quad sensor means, and even talk about the importance of depth maps.</p><p>Episode timeline:</p><p>00:00 Intro<br>02:59 iPhone 15 Overview<br>05:15 iPhone 15 Main Camera<br>07:20 Quad Pixel Sensor Explained<br>15:45 Depth Maps Explained<br>22:57 iPhone 15 Pro Overview<br>27:01 iPhone 15 Pro Cameras<br>32:20 Spatial Video<br>36:00 A17 Pro Chipset</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Mon, 18 Sep 2023 13:19:39 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/34856bf3/e10ecf64.mp3" length="101485354" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>2537</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into our in-house computer vision expert's reaction to the iPhone 15 and iPhone 15 Pro announcement.</p><p>We dive into the camera upgrades, decode what a quad sensor means, and even talk about the importance of depth maps.</p><p>Episode timeline:</p><p>00:00 Intro<br>02:59 iPhone 15 Overview<br>05:15 iPhone 15 Main Camera<br>07:20 Quad Pixel Sensor Explained<br>15:45 Depth Maps Explained<br>22:57 iPhone 15 Pro Overview<br>27:01 iPhone 15 Pro Cameras<br>32:20 Spatial Video<br>36:00 A17 Pro Chipset</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>OpenMVG Decoded: Pierre Moulon's 10 Year Journey Building Open-Source Software</title>
      <itunes:episode>10</itunes:episode>
      <podcast:episode>10</podcast:episode>
      <itunes:title>OpenMVG Decoded: Pierre Moulon's 10 Year Journey Building Open-Source Software</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">fa5c93fe-3e84-45b9-b8a5-0a0afe6f1488</guid>
      <link>https://share.transistor.fm/s/565a799f</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into Pierre Moulon's 10 years experience building OpenMVG. We also cover the impact of open-source software in the computer vision industry and everything involved in building your own project. There is a lot to learn here!</p><p>Our episode guest, Pierre Moulon, is a computer vision research scientist and creator of <a href="https://github.com/openMVG/openMVG">OpenMVG</a> -  a library for computer-vision scientists and targeted for the Multiple View Geometry community.</p><p>The episode follow's Pierre's journey building OpenMVG which he wrote about as an article in his GitHub repository.</p><p>Explore OpenMVG on GitHub: <a href="https://github.com/openMVG/openMVG">https://github.com/openMVG/openMVG</a> <br>Pierre's article on building OpenMVG: <a href="https://github.com/openMVG/openMVG/discussions/2165">https://github.com/openMVG/openMVG/discussions/2165</a></p><p>Episode timeline:</p><p>00:00 Intro<br>01:00 Pierre Moulon's Background<br>04:40 What is OpenMVG?<br>08:43 What is the importance of open-source software for the computer vision community?<br>12:30 What to look for deciding to use an opensource project<br>16:27 What is Multi View Geometry?<br>24:24 What was the biggest challenge building OpenMVG?<br>31:00 How do you grow a community around an open-source project<br>38:09 Choosing a licensing model for your open-source project<br>43:07 Funding and sponsorship for your open-source project<br>46:46 Building an open-source project for your resume<br>49:53 How to get started with OpenMVG</p><p>Contact:<br>Follow Pierre Moulon on LinkedIn: <a href="https://www.google.com/url?q=https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/&amp;sa=D&amp;source=docs&amp;ust=1682095042153388&amp;usg=AOvVaw326cbQ532ETUKgALAvS1PB">https://www.linkedin.com/in/pierre-moulon/</a><br>Follow Jared Heinly on Twitter: <a href="https://twitter.com/JaredHeinly">https://twitter.com/JaredHeinly</a><br>Follow Jonathan Stephens on Twitter at: <a href="https://twitter.com/jonstephens85">https://twitter.com/jonstephens85</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: <a href="https://www.everypoint.io/">https://www.everypoint.io</a></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into Pierre Moulon's 10 years experience building OpenMVG. We also cover the impact of open-source software in the computer vision industry and everything involved in building your own project. There is a lot to learn here!</p><p>Our episode guest, Pierre Moulon, is a computer vision research scientist and creator of <a href="https://github.com/openMVG/openMVG">OpenMVG</a> -  a library for computer-vision scientists and targeted for the Multiple View Geometry community.</p><p>The episode follow's Pierre's journey building OpenMVG which he wrote about as an article in his GitHub repository.</p><p>Explore OpenMVG on GitHub: <a href="https://github.com/openMVG/openMVG">https://github.com/openMVG/openMVG</a> <br>Pierre's article on building OpenMVG: <a href="https://github.com/openMVG/openMVG/discussions/2165">https://github.com/openMVG/openMVG/discussions/2165</a></p><p>Episode timeline:</p><p>00:00 Intro<br>01:00 Pierre Moulon's Background<br>04:40 What is OpenMVG?<br>08:43 What is the importance of open-source software for the computer vision community?<br>12:30 What to look for deciding to use an opensource project<br>16:27 What is Multi View Geometry?<br>24:24 What was the biggest challenge building OpenMVG?<br>31:00 How do you grow a community around an open-source project<br>38:09 Choosing a licensing model for your open-source project<br>43:07 Funding and sponsorship for your open-source project<br>46:46 Building an open-source project for your resume<br>49:53 How to get started with OpenMVG</p><p>Contact:<br>Follow Pierre Moulon on LinkedIn: <a href="https://www.google.com/url?q=https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/&amp;sa=D&amp;source=docs&amp;ust=1682095042153388&amp;usg=AOvVaw326cbQ532ETUKgALAvS1PB">https://www.linkedin.com/in/pierre-moulon/</a><br>Follow Jared Heinly on Twitter: <a href="https://twitter.com/JaredHeinly">https://twitter.com/JaredHeinly</a><br>Follow Jonathan Stephens on Twitter at: <a href="https://twitter.com/jonstephens85">https://twitter.com/jonstephens85</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: <a href="https://www.everypoint.io/">https://www.everypoint.io</a></p>]]>
      </content:encoded>
      <pubDate>Fri, 05 May 2023 16:08:41 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/565a799f/e0e3e670.mp3" length="80287760" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3344</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into Pierre Moulon's 10 years experience building OpenMVG. We also cover the impact of open-source software in the computer vision industry and everything involved in building your own project. There is a lot to learn here!</p><p>Our episode guest, Pierre Moulon, is a computer vision research scientist and creator of <a href="https://github.com/openMVG/openMVG">OpenMVG</a> -  a library for computer-vision scientists and targeted for the Multiple View Geometry community.</p><p>The episode follow's Pierre's journey building OpenMVG which he wrote about as an article in his GitHub repository.</p><p>Explore OpenMVG on GitHub: <a href="https://github.com/openMVG/openMVG">https://github.com/openMVG/openMVG</a> <br>Pierre's article on building OpenMVG: <a href="https://github.com/openMVG/openMVG/discussions/2165">https://github.com/openMVG/openMVG/discussions/2165</a></p><p>Episode timeline:</p><p>00:00 Intro<br>01:00 Pierre Moulon's Background<br>04:40 What is OpenMVG?<br>08:43 What is the importance of open-source software for the computer vision community?<br>12:30 What to look for deciding to use an opensource project<br>16:27 What is Multi View Geometry?<br>24:24 What was the biggest challenge building OpenMVG?<br>31:00 How do you grow a community around an open-source project<br>38:09 Choosing a licensing model for your open-source project<br>43:07 Funding and sponsorship for your open-source project<br>46:46 Building an open-source project for your resume<br>49:53 How to get started with OpenMVG</p><p>Contact:<br>Follow Pierre Moulon on LinkedIn: <a href="https://www.google.com/url?q=https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/&amp;sa=D&amp;source=docs&amp;ust=1682095042153388&amp;usg=AOvVaw326cbQ532ETUKgALAvS1PB">https://www.linkedin.com/in/pierre-moulon/</a><br>Follow Jared Heinly on Twitter: <a href="https://twitter.com/JaredHeinly">https://twitter.com/JaredHeinly</a><br>Follow Jonathan Stephens on Twitter at: <a href="https://twitter.com/jonstephens85">https://twitter.com/jonstephens85</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: <a href="https://www.everypoint.io/">https://www.everypoint.io</a></p>]]>
      </itunes:summary>
      <itunes:keywords>Computer vision, structure from motion, multi view geometry, 3d, photogrammetry, 3d reconstruction</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:person role="Guest" href="https://github.com/openMVG/openMVG" img="https://img.transistor.fm/K-_Awrwx5_NAv3SzhtQujm0wd12cuMwYqN8dewiMC-A/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZjRhNjUzMWMt/YTczNy00ZDYxLWJh/NWYtNDBjNjM4MDcw/ZDAxLzE2ODMzMjgw/MjktaW1hZ2UuanBn.jpg">Pierre Moulon</podcast:person>
    </item>
    <item>
      <title>Understanding Implicit Neural Representations with Itzik Ben-Shabat</title>
      <itunes:episode>9</itunes:episode>
      <podcast:episode>9</podcast:episode>
      <itunes:title>Understanding Implicit Neural Representations with Itzik Ben-Shabat</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">2899a5c3-4e57-4e94-a33e-2aeaea88b93d</guid>
      <link>https://share.transistor.fm/s/ee045682</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into implicit neural representations.</p><p>We are joined by Itzik Ben-Shabat, a Visiting Research Fellow at the <a href="https://www.anu.edu.au/">Australian National Universit (ANU)</a> and <a href="http://www.technion.ac.il/en/">Technion – Israel Institute of Technology</a> as well as the host of the <a href="https://talking.papers.podcast.itzikbs.com/">Talking Paper Podcast</a>.</p><p>You will learn a core understanding of implicit neural representations, key concepts and terminology, how it's being used in applications today, and Itzik's research into improving output with limit input data.</p><p>Episode timeline:</p><p>00:00 Intro<br>01:23 Overview of what implicit neural representations are<br>04:08 How INR compares and contrasts with a NeRF<br>08:17 Why did Itzik pursued this line of research<br>10:56 What is normalization and what are normals <br>13:13 Past research people should read to learn about the basics of INR <br>16:10 What is an implicit representation (without the neural network)<br>24:27 What is DiGS and what problem with INR does it solve?<br>35:54 What is OG-I NR and what problem with INR does it solve?<br>40:43 What software can researchers use to understand INR?<br>49:15 What information should non-scientists be focused to learn about INR?</p><p>Itzik's Website: <a href="https://www.google.com/url?q=https://www.itzikbs.com/&amp;sa=D&amp;source=docs&amp;ust=1682095042153130&amp;usg=AOvVaw2rinB0HhfVr3uZmCXG-BOY">https://www.itzikbs.com/</a><br>Follow Itzik on Twitter: <a href="https://www.google.com/url?q=https://twitter.com/sitzikbs&amp;sa=D&amp;source=docs&amp;ust=1682095042153263&amp;usg=AOvVaw2zp81bEQ69BLeQT1l9jS_0">https://twitter.com/sitzikbs</a><br>Follow Itzik on LinkedIn: <a href="https://www.google.com/url?q=https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/&amp;sa=D&amp;source=docs&amp;ust=1682095042153388&amp;usg=AOvVaw326cbQ532ETUKgALAvS1PB">https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/</a><br>Talking Papers Podcast: <a href="https://www.google.com/url?q=https://talking.papers.podcast.itzikbs.com/&amp;sa=D&amp;source=docs&amp;ust=1682095042153328&amp;usg=AOvVaw2fiiYgfng_9Ny3M1lIGeeT">https://talking.papers.podcast.itzikbs.com/</a></p><p>Follow Jared Heinly on Twitter: <a href="https://twitter.com/JaredHeinly">https://twitter.com/JaredHeinly</a><br>Follow Jonathan Stephens on Twitter at: <a href="https://twitter.com/jonstephens85">https://twitter.com/jonstephens85</a></p><p>Referenced past episode- What is CVPR: <a href="https://share.transistor.fm/s/15edb19d">https://share.transistor.fm/s/15edb19d</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: <a href="https://www.everypoint.io">https://www.everypoint.io</a></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into implicit neural representations.</p><p>We are joined by Itzik Ben-Shabat, a Visiting Research Fellow at the <a href="https://www.anu.edu.au/">Australian National Universit (ANU)</a> and <a href="http://www.technion.ac.il/en/">Technion – Israel Institute of Technology</a> as well as the host of the <a href="https://talking.papers.podcast.itzikbs.com/">Talking Paper Podcast</a>.</p><p>You will learn a core understanding of implicit neural representations, key concepts and terminology, how it's being used in applications today, and Itzik's research into improving output with limit input data.</p><p>Episode timeline:</p><p>00:00 Intro<br>01:23 Overview of what implicit neural representations are<br>04:08 How INR compares and contrasts with a NeRF<br>08:17 Why did Itzik pursued this line of research<br>10:56 What is normalization and what are normals <br>13:13 Past research people should read to learn about the basics of INR <br>16:10 What is an implicit representation (without the neural network)<br>24:27 What is DiGS and what problem with INR does it solve?<br>35:54 What is OG-I NR and what problem with INR does it solve?<br>40:43 What software can researchers use to understand INR?<br>49:15 What information should non-scientists be focused to learn about INR?</p><p>Itzik's Website: <a href="https://www.google.com/url?q=https://www.itzikbs.com/&amp;sa=D&amp;source=docs&amp;ust=1682095042153130&amp;usg=AOvVaw2rinB0HhfVr3uZmCXG-BOY">https://www.itzikbs.com/</a><br>Follow Itzik on Twitter: <a href="https://www.google.com/url?q=https://twitter.com/sitzikbs&amp;sa=D&amp;source=docs&amp;ust=1682095042153263&amp;usg=AOvVaw2zp81bEQ69BLeQT1l9jS_0">https://twitter.com/sitzikbs</a><br>Follow Itzik on LinkedIn: <a href="https://www.google.com/url?q=https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/&amp;sa=D&amp;source=docs&amp;ust=1682095042153388&amp;usg=AOvVaw326cbQ532ETUKgALAvS1PB">https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/</a><br>Talking Papers Podcast: <a href="https://www.google.com/url?q=https://talking.papers.podcast.itzikbs.com/&amp;sa=D&amp;source=docs&amp;ust=1682095042153328&amp;usg=AOvVaw2fiiYgfng_9Ny3M1lIGeeT">https://talking.papers.podcast.itzikbs.com/</a></p><p>Follow Jared Heinly on Twitter: <a href="https://twitter.com/JaredHeinly">https://twitter.com/JaredHeinly</a><br>Follow Jonathan Stephens on Twitter at: <a href="https://twitter.com/jonstephens85">https://twitter.com/jonstephens85</a></p><p>Referenced past episode- What is CVPR: <a href="https://share.transistor.fm/s/15edb19d">https://share.transistor.fm/s/15edb19d</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: <a href="https://www.everypoint.io">https://www.everypoint.io</a></p>]]>
      </content:encoded>
      <pubDate>Fri, 21 Apr 2023 12:42:02 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/ee045682/6d77a1db.mp3" length="79751525" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3322</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into implicit neural representations.</p><p>We are joined by Itzik Ben-Shabat, a Visiting Research Fellow at the <a href="https://www.anu.edu.au/">Australian National Universit (ANU)</a> and <a href="http://www.technion.ac.il/en/">Technion – Israel Institute of Technology</a> as well as the host of the <a href="https://talking.papers.podcast.itzikbs.com/">Talking Paper Podcast</a>.</p><p>You will learn a core understanding of implicit neural representations, key concepts and terminology, how it's being used in applications today, and Itzik's research into improving output with limit input data.</p><p>Episode timeline:</p><p>00:00 Intro<br>01:23 Overview of what implicit neural representations are<br>04:08 How INR compares and contrasts with a NeRF<br>08:17 Why did Itzik pursued this line of research<br>10:56 What is normalization and what are normals <br>13:13 Past research people should read to learn about the basics of INR <br>16:10 What is an implicit representation (without the neural network)<br>24:27 What is DiGS and what problem with INR does it solve?<br>35:54 What is OG-I NR and what problem with INR does it solve?<br>40:43 What software can researchers use to understand INR?<br>49:15 What information should non-scientists be focused to learn about INR?</p><p>Itzik's Website: <a href="https://www.google.com/url?q=https://www.itzikbs.com/&amp;sa=D&amp;source=docs&amp;ust=1682095042153130&amp;usg=AOvVaw2rinB0HhfVr3uZmCXG-BOY">https://www.itzikbs.com/</a><br>Follow Itzik on Twitter: <a href="https://www.google.com/url?q=https://twitter.com/sitzikbs&amp;sa=D&amp;source=docs&amp;ust=1682095042153263&amp;usg=AOvVaw2zp81bEQ69BLeQT1l9jS_0">https://twitter.com/sitzikbs</a><br>Follow Itzik on LinkedIn: <a href="https://www.google.com/url?q=https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/&amp;sa=D&amp;source=docs&amp;ust=1682095042153388&amp;usg=AOvVaw326cbQ532ETUKgALAvS1PB">https://www.linkedin.com/in/yizhak-itzik-ben-shabat-67b3b1b7/</a><br>Talking Papers Podcast: <a href="https://www.google.com/url?q=https://talking.papers.podcast.itzikbs.com/&amp;sa=D&amp;source=docs&amp;ust=1682095042153328&amp;usg=AOvVaw2fiiYgfng_9Ny3M1lIGeeT">https://talking.papers.podcast.itzikbs.com/</a></p><p>Follow Jared Heinly on Twitter: <a href="https://twitter.com/JaredHeinly">https://twitter.com/JaredHeinly</a><br>Follow Jonathan Stephens on Twitter at: <a href="https://twitter.com/jonstephens85">https://twitter.com/jonstephens85</a></p><p>Referenced past episode- What is CVPR: <a href="https://share.transistor.fm/s/15edb19d">https://share.transistor.fm/s/15edb19d</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: <a href="https://www.everypoint.io">https://www.everypoint.io</a></p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:person role="Guest" href="http://www.itzikbs.com/" img="https://img.transistor.fm/lA5N-D_IGOITuQUuf0Bg-tVGtlqlYs9Esz2F9wxfZNQ/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vNWVmZjdlOWYt/NmFkNy00NmU5LThj/MWUtYjQyOTY5YTRj/ZDFlLzE2ODIxMDYx/NjctaW1hZ2UuanBn.jpg">Yizhak Ben-Shabat (Itzik)</podcast:person>
    </item>
    <item>
      <title>From 2D to 3D: 4 Ways to Make a 3D Reconstruction from Imagery</title>
      <itunes:episode>8</itunes:episode>
      <podcast:episode>8</podcast:episode>
      <itunes:title>From 2D to 3D: 4 Ways to Make a 3D Reconstruction from Imagery</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">2f190157-a1b6-4810-ab27-995eedbc6ea4</guid>
      <link>https://share.transistor.fm/s/998edfb7</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into 4 different ways to 3D reconstruct a scene with images. Our cohost Jared Heinly, a PhD in the computer science specializing in 3D reconstruction from images, will dive into the 4 distinct strategies and discuss the pros and cons of each.</p><p>Links to content shared in this episode:</p><p>Live SLAM to measure a stockpile with SR Measure: https://srmeasure.com/professional</p><p>Jared's notes on the iPhone LiDAR and SLAM: https://everypoint.medium.com/everypoint-gets-hands-on-with-apples-new-lidar-sensor-44eeb38db579</p><p>How to capture images for 3D reconstruction: https://youtu.be/AQfRdr_gZ8g</p><p>00:00 Intro<br>01:30 3D Reconstruction from Video<br>13:48 3D Reconstruction from Images<br>28:05 3D Reconstruction from Stereo Pairs<br>38:43 3D Reconstruction from SLAM</p><p>Follow Jared Heinly <br>Twitter: https://twitter.com/JaredHeinly<br>LinkedIn https://www.linkedin.com/in/jheinly/</p><p>Follow Jonathan Stephens<br>Twitter: https://twitter.com/jonstephens85<br>LinkedIn: https://www.linkedin.com/in/jonathanstephens/</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into 4 different ways to 3D reconstruct a scene with images. Our cohost Jared Heinly, a PhD in the computer science specializing in 3D reconstruction from images, will dive into the 4 distinct strategies and discuss the pros and cons of each.</p><p>Links to content shared in this episode:</p><p>Live SLAM to measure a stockpile with SR Measure: https://srmeasure.com/professional</p><p>Jared's notes on the iPhone LiDAR and SLAM: https://everypoint.medium.com/everypoint-gets-hands-on-with-apples-new-lidar-sensor-44eeb38db579</p><p>How to capture images for 3D reconstruction: https://youtu.be/AQfRdr_gZ8g</p><p>00:00 Intro<br>01:30 3D Reconstruction from Video<br>13:48 3D Reconstruction from Images<br>28:05 3D Reconstruction from Stereo Pairs<br>38:43 3D Reconstruction from SLAM</p><p>Follow Jared Heinly <br>Twitter: https://twitter.com/JaredHeinly<br>LinkedIn https://www.linkedin.com/in/jheinly/</p><p>Follow Jonathan Stephens<br>Twitter: https://twitter.com/jonstephens85<br>LinkedIn: https://www.linkedin.com/in/jonathanstephens/</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Thu, 16 Mar 2023 09:47:58 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/998edfb7/e09849d9.mp3" length="78479353" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3269</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into 4 different ways to 3D reconstruct a scene with images. Our cohost Jared Heinly, a PhD in the computer science specializing in 3D reconstruction from images, will dive into the 4 distinct strategies and discuss the pros and cons of each.</p><p>Links to content shared in this episode:</p><p>Live SLAM to measure a stockpile with SR Measure: https://srmeasure.com/professional</p><p>Jared's notes on the iPhone LiDAR and SLAM: https://everypoint.medium.com/everypoint-gets-hands-on-with-apples-new-lidar-sensor-44eeb38db579</p><p>How to capture images for 3D reconstruction: https://youtu.be/AQfRdr_gZ8g</p><p>00:00 Intro<br>01:30 3D Reconstruction from Video<br>13:48 3D Reconstruction from Images<br>28:05 3D Reconstruction from Stereo Pairs<br>38:43 3D Reconstruction from SLAM</p><p>Follow Jared Heinly <br>Twitter: https://twitter.com/JaredHeinly<br>LinkedIn https://www.linkedin.com/in/jheinly/</p><p>Follow Jonathan Stephens<br>Twitter: https://twitter.com/jonstephens85<br>LinkedIn: https://www.linkedin.com/in/jonathanstephens/</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </itunes:summary>
      <itunes:keywords>computer vision, machine learning, apple, photogrammetry, spatial computing.</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>From Concept to Reality: The Journey of Building Scaniverse</title>
      <itunes:episode>7</itunes:episode>
      <podcast:episode>7</podcast:episode>
      <itunes:title>From Concept to Reality: The Journey of Building Scaniverse</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">59e74a5c-d7f4-45c1-bc05-b19492b67b41</guid>
      <link>https://share.transistor.fm/s/8e98e86e</link>
      <description>
        <![CDATA[<p>Join our guest, Keith Ito, founder of Scaniverse as we discuss the challenges of creating a 3D capture app for iPhones. Keith goes into depth on balancing speed with quality of 3D output and how he designed an intuitive user experience for his users.</p><p><em>In this episode, we discuss…</em></p><ul><li>01:00 - Keith's Ito's background at Google</li><li>09:44 - What is the Scaniverse app</li><li>11:43 - What inspired Keith to build Scaniverse</li><li>17:37 - The challenges of using LiDAR in the early versions of Scaniverse</li><li>25:54 - How to build a good user experience for 3D capture apps</li><li>32:00 - The challenges of running photogrammetry on an iPhone</li><li>37:07 - The future of 3D capture</li><li>40:57 - Scaniverse's role at Niantic</li></ul><p>Learn more about Scaniverse at: https://scaniverse.com/<br>Follow Keith Ito on Twitter at: https://twitter.com/keeeto</p><p>Follow Jared Heinly on Twitter: https://twitter.com/JaredHeinly<br>Follow Jonathan Stephens on Twitter: https://twitter.com/jonstephens85<br>Follow Jonathan Stephens on LinkedIn: https://www.linkedin.com/in/jonathanstephens/</p><p>-----</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>Join our guest, Keith Ito, founder of Scaniverse as we discuss the challenges of creating a 3D capture app for iPhones. Keith goes into depth on balancing speed with quality of 3D output and how he designed an intuitive user experience for his users.</p><p><em>In this episode, we discuss…</em></p><ul><li>01:00 - Keith's Ito's background at Google</li><li>09:44 - What is the Scaniverse app</li><li>11:43 - What inspired Keith to build Scaniverse</li><li>17:37 - The challenges of using LiDAR in the early versions of Scaniverse</li><li>25:54 - How to build a good user experience for 3D capture apps</li><li>32:00 - The challenges of running photogrammetry on an iPhone</li><li>37:07 - The future of 3D capture</li><li>40:57 - Scaniverse's role at Niantic</li></ul><p>Learn more about Scaniverse at: https://scaniverse.com/<br>Follow Keith Ito on Twitter at: https://twitter.com/keeeto</p><p>Follow Jared Heinly on Twitter: https://twitter.com/JaredHeinly<br>Follow Jonathan Stephens on Twitter: https://twitter.com/jonstephens85<br>Follow Jonathan Stephens on LinkedIn: https://www.linkedin.com/in/jonathanstephens/</p><p>-----</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Tue, 24 Jan 2023 12:39:00 -0800</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/8e98e86e/f06b1fca.mp3" length="72145165" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3005</itunes:duration>
      <itunes:summary>Join our guest, Keith Ito, founder of Scaniverse as we discuss the challenges of creating a 3D capture app for iPhones. Keith goes into depth on balancing speed with quality of 3D output and how he designed an intuitive user experience for his users.</itunes:summary>
      <itunes:subtitle>Join our guest, Keith Ito, founder of Scaniverse as we discuss the challenges of creating a 3D capture app for iPhones. Keith goes into depth on balancing speed with quality of 3D output and how he designed an intuitive user experience for his users.</itunes:subtitle>
      <itunes:keywords>computer vision, lidar, photogrammetry, iphone app, startup</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:person role="Guest" href="https://keithito.com" img="https://img.transistor.fm/jX-eb4KBDtS6HeRVgcmSYaugvSCsAuVLedI7JsQatC4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOGI3MGYwMDEt/MzI1YS00MGI3LWE5/YTEtMjIwNDc3YWFh/NGQ4LzE2NzQ1MTkx/NDctaW1hZ2UuanBn.jpg">Keith Ito</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/8e98e86e/transcript.txt" type="text/plain"/>
    </item>
    <item>
      <title>Will NeRFs Replace Photogrammetry?</title>
      <itunes:episode>6</itunes:episode>
      <podcast:episode>6</podcast:episode>
      <itunes:title>Will NeRFs Replace Photogrammetry?</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">bb4f8d31-3eb3-425c-9fb4-29ee7d199199</guid>
      <link>https://share.transistor.fm/s/3ac132a4</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into one of the hottest topics in the industry: Neural Radiance Fields (NeRFs)</p><p>We are joined by Matt Tancik, a student pursuing a PhD in the computer science and electrical engineering department at UC Berkeley. He has also contributed research to the original NeRF project in 2020 along with several others since then.</p><p>Last but not least, he is building NeRFStudio - a collaboration friendly studio for NeRFs.</p><p>In this episode you will learn about what NeRFs are and more importantly what they are not. Matt goes into the challenges of large scale NeRF creation with his experience with Block-NeRF.</p><p>Follow Matt's work at https://www.matthewtancik.com/</p><p>Get started with Nerfstudio here: https://docs.nerf.studio/en/latest/</p><p>Block-NeRF details: https://waymo.com/research/block-nerf/</p><p>00:00 Intro<br>00:45 Matt’s Background Into NeRF Research <br>04:00 What is a NeRF and how it is different from photogrammetry<br>11:57 Can geometry be extracted from NeRFs?<br>15:30 Will NeRFs supersede photogrammetry in the future? <br>22:47 Block-NeRF and the pros and cons of using 360 cameras<br>25:30 What is the goal of Block-NeRF<br>30:44 Why do NeRFs need large GPUs to compute?<br>35:45 Meshes to simulate NeRF visualizations<br>40:28 What is Nerfstudio?<br>47:40 How to get started with Nerfstudio</p><p>Follow Jared Heinly on Twitter: https://twitter.com/JaredHeinly<br>Follow Jonathan Stephens on Twitter at: https://twitter.com/jonstephens85</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into one of the hottest topics in the industry: Neural Radiance Fields (NeRFs)</p><p>We are joined by Matt Tancik, a student pursuing a PhD in the computer science and electrical engineering department at UC Berkeley. He has also contributed research to the original NeRF project in 2020 along with several others since then.</p><p>Last but not least, he is building NeRFStudio - a collaboration friendly studio for NeRFs.</p><p>In this episode you will learn about what NeRFs are and more importantly what they are not. Matt goes into the challenges of large scale NeRF creation with his experience with Block-NeRF.</p><p>Follow Matt's work at https://www.matthewtancik.com/</p><p>Get started with Nerfstudio here: https://docs.nerf.studio/en/latest/</p><p>Block-NeRF details: https://waymo.com/research/block-nerf/</p><p>00:00 Intro<br>00:45 Matt’s Background Into NeRF Research <br>04:00 What is a NeRF and how it is different from photogrammetry<br>11:57 Can geometry be extracted from NeRFs?<br>15:30 Will NeRFs supersede photogrammetry in the future? <br>22:47 Block-NeRF and the pros and cons of using 360 cameras<br>25:30 What is the goal of Block-NeRF<br>30:44 Why do NeRFs need large GPUs to compute?<br>35:45 Meshes to simulate NeRF visualizations<br>40:28 What is Nerfstudio?<br>47:40 How to get started with Nerfstudio</p><p>Follow Jared Heinly on Twitter: https://twitter.com/JaredHeinly<br>Follow Jonathan Stephens on Twitter at: https://twitter.com/jonstephens85</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Fri, 11 Nov 2022 12:25:42 -0800</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/3ac132a4/1f72514f.mp3" length="75228648" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3134</itunes:duration>
      <itunes:summary>In this episode of Computer Vision Decoded, we are going to dive into one of the hottest topics in the industry: Neural Radiance Fields (NeRFs)

We are joined by Matt Tancik, a student pursuing a PhD in the computer science and electrical engineering department at UC Berkeley. He has also contributed research to the original NeRF project in 2020 along with several others since then.

Last but not least, he is building NeRFStudio - a collaboration friendly studio for NeRFs.

In this episode you will learn about what NeRFs are and more importantly what they are not. Matt goes into the challenges of large scale NeRF creation with his experience with Block-NeRF.</itunes:summary>
      <itunes:subtitle>In this episode of Computer Vision Decoded, we are going to dive into one of the hottest topics in the industry: Neural Radiance Fields (NeRFs)

We are joined by Matt Tancik, a student pursuing a PhD in the computer science and electrical engineering de</itunes:subtitle>
      <itunes:keywords>Neural radiance fields, NeRF, computer vision, 3d, AI, machine learning</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>How to Capture Images for 3D Reconstruction</title>
      <itunes:episode>5</itunes:episode>
      <podcast:episode>5</podcast:episode>
      <itunes:title>How to Capture Images for 3D Reconstruction</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">d800dbd0-e6c6-428f-98f5-42e2ea41b0cb</guid>
      <link>https://share.transistor.fm/s/34066f83</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into image capture best practices for 3D reconstruction.</p><p>At the end of this livestream, you will have learned the basics for capturing scenes and objects. We will also provide a downloadable visual guide for reference on your next 3D reconstruction project.</p><p>Download the official guide here to follow along: https://tinyurl.com/4n2wspkn</p><p>00:00 Intro<br>04:40 Camera motion overview<br>07:15 Good camera motions<br>18:43 Transition camera motions<br>30:39 Bad camera motions<br>39:27 How to combine camera motions<br>49:16 Loop Closure<br>57:42 Image Overlap<br>1:14:00 Lighting and camera gear</p><p>Watch out episode of Computer Vision in the Wild to learn more about capturing images outside and in busy locations: https://youtu.be/FwVBR6KFjPI</p><p>Follow Jared Heinly on Twitter: https://twitter.com/JaredHeinly<br>Follow Jonathan Stephens on Twitter at: https://twitter.com/jonstephens85</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we are going to dive into image capture best practices for 3D reconstruction.</p><p>At the end of this livestream, you will have learned the basics for capturing scenes and objects. We will also provide a downloadable visual guide for reference on your next 3D reconstruction project.</p><p>Download the official guide here to follow along: https://tinyurl.com/4n2wspkn</p><p>00:00 Intro<br>04:40 Camera motion overview<br>07:15 Good camera motions<br>18:43 Transition camera motions<br>30:39 Bad camera motions<br>39:27 How to combine camera motions<br>49:16 Loop Closure<br>57:42 Image Overlap<br>1:14:00 Lighting and camera gear</p><p>Watch out episode of Computer Vision in the Wild to learn more about capturing images outside and in busy locations: https://youtu.be/FwVBR6KFjPI</p><p>Follow Jared Heinly on Twitter: https://twitter.com/JaredHeinly<br>Follow Jonathan Stephens on Twitter at: https://twitter.com/jonstephens85</p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Fri, 23 Sep 2022 22:06:31 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/34066f83/7fcbc52d.mp3" length="80170741" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>5009</itunes:duration>
      <itunes:summary>In this episode of Computer Vision Decoded, we are going to dive into image capture best practices for 3D reconstruction.</itunes:summary>
      <itunes:subtitle>In this episode of Computer Vision Decoded, we are going to dive into image capture best practices for 3D reconstruction.</itunes:subtitle>
      <itunes:keywords>computer vision, photogrammetry, 3d, 3d reconstruction, computer science</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>Is The iPhone 14 Camera Any Good?</title>
      <itunes:episode>4</itunes:episode>
      <podcast:episode>4</podcast:episode>
      <itunes:title>Is The iPhone 14 Camera Any Good?</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">32a03a78-c9e8-451c-abe0-00daa5cfd28a</guid>
      <link>https://share.transistor.fm/s/8a0ffe8e</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we join Jared Heinly and Jonathan Stephens from EveryPoint for their live reaction to the iPhone 14 series announcement. They go in depth into what all the camera specs mean to the average person. We also explain basics of computational photography and how Apple is able to get great photos from a small camera sensor.</p><p>00:00 Intro<br>02:43 Apple Watch Review<br>06:58 Airpods Pro Review <br>09:40 iPhone 14 Initial Reaction<br>15:05 iPhone 14 Camera Specs Breakdown<br>37:13 iPhone 14 Pro Initial Reaction<br>40:47 iPhone 14 Pro Camera Specs Breakdown</p><p><a href="https://twitter.com/JaredHeinly">Follow Jared Heinly on Twitter</a><br><a href="https://twitter.com/jonstephens85">Follow Jonathan Stephens on Twitter</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we join Jared Heinly and Jonathan Stephens from EveryPoint for their live reaction to the iPhone 14 series announcement. They go in depth into what all the camera specs mean to the average person. We also explain basics of computational photography and how Apple is able to get great photos from a small camera sensor.</p><p>00:00 Intro<br>02:43 Apple Watch Review<br>06:58 Airpods Pro Review <br>09:40 iPhone 14 Initial Reaction<br>15:05 iPhone 14 Camera Specs Breakdown<br>37:13 iPhone 14 Pro Initial Reaction<br>40:47 iPhone 14 Pro Camera Specs Breakdown</p><p><a href="https://twitter.com/JaredHeinly">Follow Jared Heinly on Twitter</a><br><a href="https://twitter.com/jonstephens85">Follow Jonathan Stephens on Twitter</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Thu, 08 Sep 2022 11:42:50 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/8a0ffe8e/e4dd2657.mp3" length="88670018" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3694</itunes:duration>
      <itunes:summary>In this episode of computer vision decoded we recorded our live reaction to the iPhone 14 series announcement. We go in depth into what all the camera specs mean to the average person. We also explain basics of computational photography and how Apple is able to get great photos from a small camera sensor.</itunes:summary>
      <itunes:subtitle>In this episode of computer vision decoded we recorded our live reaction to the iPhone 14 series announcement. We go in depth into what all the camera specs mean to the average person. We also explain basics of computational photography and how Apple is a</itunes:subtitle>
      <itunes:keywords>iphone, apple, camera, computer vision, computational photography, photography</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>3D Reconstruction in the Wild</title>
      <itunes:episode>3</itunes:episode>
      <podcast:episode>3</podcast:episode>
      <itunes:title>3D Reconstruction in the Wild</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">5354aa69-6aaf-48d3-a51c-404c38e6ff5e</guid>
      <link>https://share.transistor.fm/s/7740b1ed</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we sit down with Jared Heinly, Chief Scientist at EveryPoint, to discuss 3D reconstruction in the wild. What does “in the wild” mean? This means 3D reconstructing objects and scenes in non-controlled environments where you may have limitations with lighting, access, reflective surfaces, etc.</p><p>00:00 Intro<br>01:30: What are Duplicate Scene Structures and How to Avoid Them<br>14:30: How Jared used 100 million crowdsourced photos to 3d reconstruct 12,903 landmarks<br>27:10: The benefits of capturing video for 3D reconstruction<br>31:30: The benefits of using a drone to capture stills for 3D reconstruction<br>34:20: Considerations for using installed cameras for 3d reconstruction<br>38:30: How to work with sun issues<br>44:25: Determining how far from the object you should be when capturing images<br>50:35: How to capture objects with reflective surfaces<br>53:40: How work around scene obstructions<br>57:20: What cameras you should use</p><p>Jared Heinly’s Academic Papers and Projects</p><p>Paper: <a href="https://www.cs.unc.edu/~jheinly/publications/3dv2014-heinly.pdf">Correcting the Duplicate Scene Structure In Sparse 3D Reconstruction</a><br>Project: <a href="https://www.cs.unc.edu/~jheinly/reconstructing_the_world.html">Reconstructing the World in Six Days</a><br>Video: <a href="https://youtu.be/bRYqyoqUJuM">Reconstructing the world in Six Days</a></p><p><a href="https://twitter.com/JaredHeinly">Follow Jared Heinly on Twitter</a><br><a href="https://twitter.com/jonstephens85">Follow Jonathan Stephens on Twitter</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded, we sit down with Jared Heinly, Chief Scientist at EveryPoint, to discuss 3D reconstruction in the wild. What does “in the wild” mean? This means 3D reconstructing objects and scenes in non-controlled environments where you may have limitations with lighting, access, reflective surfaces, etc.</p><p>00:00 Intro<br>01:30: What are Duplicate Scene Structures and How to Avoid Them<br>14:30: How Jared used 100 million crowdsourced photos to 3d reconstruct 12,903 landmarks<br>27:10: The benefits of capturing video for 3D reconstruction<br>31:30: The benefits of using a drone to capture stills for 3D reconstruction<br>34:20: Considerations for using installed cameras for 3d reconstruction<br>38:30: How to work with sun issues<br>44:25: Determining how far from the object you should be when capturing images<br>50:35: How to capture objects with reflective surfaces<br>53:40: How work around scene obstructions<br>57:20: What cameras you should use</p><p>Jared Heinly’s Academic Papers and Projects</p><p>Paper: <a href="https://www.cs.unc.edu/~jheinly/publications/3dv2014-heinly.pdf">Correcting the Duplicate Scene Structure In Sparse 3D Reconstruction</a><br>Project: <a href="https://www.cs.unc.edu/~jheinly/reconstructing_the_world.html">Reconstructing the World in Six Days</a><br>Video: <a href="https://youtu.be/bRYqyoqUJuM">Reconstructing the world in Six Days</a></p><p><a href="https://twitter.com/JaredHeinly">Follow Jared Heinly on Twitter</a><br><a href="https://twitter.com/jonstephens85">Follow Jonathan Stephens on Twitter</a></p><p>This episode is brought to you by EveryPoint. Learn more about how EveryPoint is building an infinitely scalable data collection and processing platform for the next generation of spatial computing applications and services: https://www.everypoint.io</p>]]>
      </content:encoded>
      <pubDate>Tue, 09 Aug 2022 14:20:07 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/7740b1ed/57430365.mp3" length="89112913" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>3711</itunes:duration>
      <itunes:summary>In this episode of Computer Vision Decoded, we sit down with Jared Heinly, Chief Scientist at EveryPoint, to discuss 3D reconstruction in the wild. What does “in the wild” mean? This means 3D reconstructing objects and scenes in non-controlled environments where you may have limitations with lighting, access, reflective surfaces, etc.</itunes:summary>
      <itunes:subtitle>In this episode of Computer Vision Decoded, we sit down with Jared Heinly, Chief Scientist at EveryPoint, to discuss 3D reconstruction in the wild. What does “in the wild” mean? This means 3D reconstructing objects and scenes in non-controlled environment</itunes:subtitle>
      <itunes:keywords>computer vision, 3d, photogrammetry, 3D Reconstruction, drones</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
    </item>
    <item>
      <title>What is the CVPR Conference?</title>
      <itunes:episode>2</itunes:episode>
      <podcast:episode>2</podcast:episode>
      <itunes:title>What is the CVPR Conference?</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">8f291f3b-1a70-4161-96af-fdf82f617e92</guid>
      <link>https://share.transistor.fm/s/15edb19d</link>
      <description>
        <![CDATA[<p>In this episode of Computer Vision Decoded we dive into Jared Heinly's recent trip to the CVPR Conference. We cover: what the conference about, who should attend, what are the emerging trends in computer vision, how machine learning is being used in 3D reconstruction, and what NeRFs are for.<em></em></p><p>00:00 - Introduction<br>00:36 - What is CVPR?<br>02:49 - Who should attend CVPR?<br>08:11 - What are emerging trends in Computer Vision?<br>14:34 - What is the value of NeRFs?<br>20:55 - How should you attend as a non-scientist or academic?</p><p><a href="https://twitter.com/JaredHeinly"><em>Follow Jared Heinly on Twitter</em></a><em><br></em><a href="https://twitter.com/jonstephens85"><em>Follow Jonathan Stephens on Twitter</em></a><em><br></em><strong><em><br></em></strong><a href="https://cvpr2022.thecvf.com/"><em>CVPR Conference</em></a><em></em></p><p><strong>Episode sponsored by: </strong><a href="https://www.everypoint.io/">EveryPoint</a></p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this episode of Computer Vision Decoded we dive into Jared Heinly's recent trip to the CVPR Conference. We cover: what the conference about, who should attend, what are the emerging trends in computer vision, how machine learning is being used in 3D reconstruction, and what NeRFs are for.<em></em></p><p>00:00 - Introduction<br>00:36 - What is CVPR?<br>02:49 - Who should attend CVPR?<br>08:11 - What are emerging trends in Computer Vision?<br>14:34 - What is the value of NeRFs?<br>20:55 - How should you attend as a non-scientist or academic?</p><p><a href="https://twitter.com/JaredHeinly"><em>Follow Jared Heinly on Twitter</em></a><em><br></em><a href="https://twitter.com/jonstephens85"><em>Follow Jonathan Stephens on Twitter</em></a><em><br></em><strong><em><br></em></strong><a href="https://cvpr2022.thecvf.com/"><em>CVPR Conference</em></a><em></em></p><p><strong>Episode sponsored by: </strong><a href="https://www.everypoint.io/">EveryPoint</a></p>]]>
      </content:encoded>
      <pubDate>Fri, 01 Jul 2022 16:19:18 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/15edb19d/08f092df.mp3" length="28154924" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>1757</itunes:duration>
      <itunes:summary>
        <![CDATA[<p>In this episode of Computer Vision Decoded we dive into Jared Heinly's recent trip to the CVPR Conference. We cover: what the conference about, who should attend, what are the emerging trends in computer vision, how machine learning is being used in 3D reconstruction, and what NeRFs are for.<em></em></p><p>00:00 - Introduction<br>00:36 - What is CVPR?<br>02:49 - Who should attend CVPR?<br>08:11 - What are emerging trends in Computer Vision?<br>14:34 - What is the value of NeRFs?<br>20:55 - How should you attend as a non-scientist or academic?</p><p><a href="https://twitter.com/JaredHeinly"><em>Follow Jared Heinly on Twitter</em></a><em><br></em><a href="https://twitter.com/jonstephens85"><em>Follow Jonathan Stephens on Twitter</em></a><em><br></em><strong><em><br></em></strong><a href="https://cvpr2022.thecvf.com/"><em>CVPR Conference</em></a><em></em></p><p><strong>Episode sponsored by: </strong><a href="https://www.everypoint.io/">EveryPoint</a></p>]]>
      </itunes:summary>
      <itunes:keywords>Computer Vision, CVPR, Machine Learning, ML, NeRF, Neural Radiance Fields, Neural Networks</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/15edb19d/transcript.txt" type="text/plain"/>
    </item>
    <item>
      <title>What Do the WWDC Announcements Mean for Computer Vision?</title>
      <itunes:episode>1</itunes:episode>
      <podcast:episode>1</podcast:episode>
      <itunes:title>What Do the WWDC Announcements Mean for Computer Vision?</itunes:title>
      <itunes:episodeType>full</itunes:episodeType>
      <guid isPermaLink="false">5e0af863-2ad7-462c-a500-e3b7e28f50f7</guid>
      <link>https://share.transistor.fm/s/48a1be0a</link>
      <description>
        <![CDATA[<p>In this inaugural episode of Computer Vision Decoded we dive into the recent announcements at WWDC 2022 and find out what they mean for the computer vision community. We talk about what Apple is doing with their new RoomPlan API and how computer vision scientists can leverage it for better experiences. We also cover the enhancements to video and photo capture during an active ARKit Session.<em></em></p><p>00:00 - Introduction<br>00:25 - Meet Jared Heinly<br>02:10 - RoomPlan API<br>06:23 - Higher Resolution Video with ARKit<br>09:17 - The importance of pixel size and density<br>13:13 - Copy and Paste Objects from Photos<br>16:47 - CVPR Conference Overview</p><p><a href="https://twitter.com/JaredHeinly"><em>Follow Jared Heinly on Twitter</em></a><em><br></em><a href="https://twitter.com/jonstephens85"><em>Follow Jonathan Stephens on Twitter</em></a><em><br></em><strong><em><br></em></strong><a href="https://developer.apple.com/augmented-reality/roomplan/"><em>Learn about RoomPlan API Overview</em></a><em><br></em><a href="https://developer.apple.com/augmented-reality/arkit/"><em>Learn about ARKit 6 Highlights</em></a><em><br></em><a href="https://cvpr2022.thecvf.com/"><em>CVPR Conference</em></a><em></em></p><p><strong>Episode sponsored by: </strong><a href="https://www.everypoint.io">EveryPoint</a> </p>]]>
      </description>
      <content:encoded>
        <![CDATA[<p>In this inaugural episode of Computer Vision Decoded we dive into the recent announcements at WWDC 2022 and find out what they mean for the computer vision community. We talk about what Apple is doing with their new RoomPlan API and how computer vision scientists can leverage it for better experiences. We also cover the enhancements to video and photo capture during an active ARKit Session.<em></em></p><p>00:00 - Introduction<br>00:25 - Meet Jared Heinly<br>02:10 - RoomPlan API<br>06:23 - Higher Resolution Video with ARKit<br>09:17 - The importance of pixel size and density<br>13:13 - Copy and Paste Objects from Photos<br>16:47 - CVPR Conference Overview</p><p><a href="https://twitter.com/JaredHeinly"><em>Follow Jared Heinly on Twitter</em></a><em><br></em><a href="https://twitter.com/jonstephens85"><em>Follow Jonathan Stephens on Twitter</em></a><em><br></em><strong><em><br></em></strong><a href="https://developer.apple.com/augmented-reality/roomplan/"><em>Learn about RoomPlan API Overview</em></a><em><br></em><a href="https://developer.apple.com/augmented-reality/arkit/"><em>Learn about ARKit 6 Highlights</em></a><em><br></em><a href="https://cvpr2022.thecvf.com/"><em>CVPR Conference</em></a><em></em></p><p><strong>Episode sponsored by: </strong><a href="https://www.everypoint.io">EveryPoint</a> </p>]]>
      </content:encoded>
      <pubDate>Tue, 21 Jun 2022 13:27:17 -0700</pubDate>
      <author>EveryPoint</author>
      <enclosure url="https://media.transistor.fm/48a1be0a/f2def4c5.mp3" length="20035027" type="audio/mpeg"/>
      <itunes:author>EveryPoint</itunes:author>
      <itunes:duration>1249</itunes:duration>
      <itunes:summary>In this episode of Computer Vision Decoded, we sit down with Jared Heinly, Chief Scientist at EveryPoint, to discuss announcements made last week at Apple’s WWDC conference and find out what they mean for the computer vision community.</itunes:summary>
      <itunes:subtitle>In this episode of Computer Vision Decoded, we sit down with Jared Heinly, Chief Scientist at EveryPoint, to discuss announcements made last week at Apple’s WWDC conference and find out what they mean for the computer vision community.</itunes:subtitle>
      <itunes:keywords>wwdc, apple, arkit, computer vision, cvpr</itunes:keywords>
      <itunes:explicit>No</itunes:explicit>
      <podcast:person role="Host" href="http://jaredheinly.com" img="https://img.transistor.fm/YPH7qWt-HWsguOoLIo9SOfKry0Wv6KJhFc--MYIloH4/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vZDZmOTk5ZDIt/NmM3Zi00MzExLWI3/Y2UtNWJhOTNhZGM1/Y2U4LzE2Njc0MzM0/NjctaW1hZ2UuanBn.jpg">Jared Heinly</podcast:person>
      <podcast:person role="Host" href="https://linktr.ee/jonstephens" img="https://img.transistor.fm/a2RByktTDbl1GOJcqi2EYEkpYHxBwGDIwC0RAY5mrcI/rs:fill:0:0:1/w:800/h:800/q:60/mb:500000/aHR0cHM6Ly9pbWct/dXBsb2FkLXByb2R1/Y3Rpb24udHJhbnNp/c3Rvci5mbS9wZXJz/b24vOWM0ZmExZDgt/ZTcxNy00OGQ5LWEy/MjgtMjUxODY4NDUw/MWI2LzE2NjkwNDcx/NjItaW1hZ2UuanBn.jpg">Jonathan Stephens</podcast:person>
      <podcast:transcript url="https://share.transistor.fm/s/48a1be0a/transcript.txt" type="text/plain"/>
    </item>
  </channel>
</rss>
