Skip to main content

Watch Google’s latest deep-learning system recognize sports in YouTube clips

Google thinks this is "mountain unicycling."
Image Credit: Screen shot

testsetset

Researchers at Google, Facebook, and other companies are working hard to use artificial intelligence to understand what’s going on in videos — and in pictures, and in speech. Today Google showed off its latest breakthroughs in research, involving a trendy type of AI called deep learning.

This approach often involves ingesting lots of data to train systems called neural networks, and then feeding new data to those systems and receiving predictions in response.

In Google’s case, researchers tested out several methods in order to correctly recognize objects and interpret motion in videos of sports: recurrent neural networks and feature-pooling networks, in combination with widely used convolutional neural networks.

“We conclude by observing that although very different in concept, the max-pooling and the recurrent neural network methods perform similarly when using both images and optical flow,” Google software engineers George Toderici and Sudheendra Vijayanarasimhan wrote in a blog post today on their work, which will be presented at the Computer Vision and Pattern Recognition conference in Boston in June.


June 5th: The AI Audit in NYC

Join us next week in NYC to engage with top executive leaders, delving into strategies for auditing AI models to ensure fairness, optimal performance, and ethical compliance across diverse organizations. Secure your attendance for this exclusive invite-only event.


You can read the academic paper, or, to get a sense of Google’s latest video-processing capabilities, you can just watch this video:

[youtube http://www.youtube.com/watch?v=oDRl3-X1KkI&w=560&h=315]