
Machine learning proves its worth for new video effects tech: distinguishing between faces and backgrounds at 100 frames per second.
![]() |
Google's AI software can detect the subject of a video and change the background. |
Switching out background scenery is a decades-old technology, but it's generally complicated and time-consuming -- think of making movies with superheros acting in front of green screens so computers can later replace the uniform green color with digital effects like exploding buildings. Having 3D scene data, for example extracted from newer iPhones' twin cameras, makes the process only a little easier.
Google, though, trained a neural network with lots of carefully labeled imagery that helped it learn how to distinguish facial features -- eyes, hair, glasses, mouths, and so on -- from everything else. The result is a system that can swap out backgrounds fast enough to keep up with video. Digital video frames whip by at 30 frames per second, but Google's technology works at 40 frames per second on a Google Pixel 2 phone and more than 100 frames per second on an Apple iPhone 7.
Want to try it out for yourself? Sorry. For now, only a pretty small number of high-profile YouTubers with access to the YouTube stories service, which offers an ability to share Snapchat-esque short videos.
Google detailed the research in a blog post Thursday.
"Our new segmentation technology allows creators to replace and modify the background, effortlessly increasing videos' production value without specialized equipment," Google programmers Valentin Bazarevsky and Andrei Tkachenka said in the blog post.
![]() |
Google uses machine learning in its video segmentation technology that can determine background imagery that can be replaced with something funny or eye-catching. |
Google already uses AI in a related feature today on its Google Pixel 2 phones. The phones only have single cameras, but AI helps blur backgrounds for portrait-mode photos.
source:CNet News
Post a Comment