What you need to know

  • Google is training its AI to create depth maps isolating human subjects in the scene using only one camera.
  • As a starting point, Google used 2000 YouTube videos from the Mannequin Challenge to train the AI.
  • The results will lead to the ability to add effects to videos, such as portrait mode, and be used for Augmented Reality.

In a recent blog post, Google detailed how it has been working on depth perception in videos where both the camera and subject are moving. As a starting point, the study needed access to a vast amount of data to train the AI, and the first logical step was training it to detect people in a scene where the camera was moving but the people were static.

As it turns out, Google had the perfect resource for this data in the form of YouTube videos that were filmed for the Mannequin Challenge. In this challenge, a person or group of people would stand completely still as a camera panned around their position. Google used 2000 videos from the challenge to help train its AI to detect human figures in a variety of different scenes.

Save big on these VPN services ahead of Black Friday

Something that makes this study even more interesting is the fact that Google is teaching its AI to create depth maps using footage that has been shot using only one camera. Most times, multiple cameras must be used to sense depth information in a scene.

Google already utilizes something similar for still images to create its portrait mode effect on the Pixel phones. However, this only pertains to still images. The new method Google has been developing is training its AI to create a depth map where both the camera and subject are moving within a scene.

By branching out into videos, it will open up features in the future for creating bokeh in video scenes similar to portrait mode on your phone. Another benefit to come from this study will be improved results for augmented reality, such as the Playmojis from Google's Playground.

Another possibility will be the generation of 3D images from 2D scenes. While camera hardware has always been essential for photography and videography, what Google has done over the years with software shows that, in the future, algorithms will be just as important and help provide new experiences.

Get More Pixel 3a

Google Pixel 3a

Have you listened to this week's Android Central Podcast?

Android Central

Every week, the Android Central Podcast brings you the latest tech news, analysis and hot takes, with familiar co-hosts and special guests.

  • Subscribe in Pocket Casts: Audio
  • Subscribe in Spotify: Audio
  • Subscribe in iTunes: Audio

We may earn a commission for purchases using our links. Learn more.