Google Brain

AI Research

GLOM: How to represent part-whole hierarchies in a neural network (Geoff Hinton’s Paper Explained)

Yannic Kilcher covers a paper where Geoffrey Hinton describes GLOM, a Computer Vision model that combines transformers, neural fields, contrastive learning, capsule networks, denoising autoencoders and RNNs. GLOM decomposes an image into a parse tree of objects and their parts. However, unlike previous systems, the parse tree is constructed dynamically and differently for each input, […]

Read More
AI Natural Language Processing

Transformers for Image Recognition at Scale

Yannic Kilcher explains why transformers are ruining convolutions. This paper, under review at ICLR, shows that given enough data, a standard Transformer can outperform Convolutional Neural Networks in image recognition tasks, which are classically tasks where CNNs excel. In this Video, I explain the architecture of the Vision Transformer (ViT), the reason why it works […]

Read More
Shark or Baseball? Inside the ‘Black Box’ of a Neural Network
AI Research

Shark or Baseball? Inside the ‘Black Box’ of a Neural Network

Generally speaking, Neural Networks are somewhat of a mystery. While you can understand the mechanics and the math that powers them, exactly how the network comes to its conclusions are a bit of a black box. Here’s an interesting story on how researchers are trying to peer into the mysteries of a neural net. Using […]

Read More
AI

AI that Dresses Itself

Siraj Raval explores how a team of researchers at Google Brain and Georgia Tech developed an AI that learned how to dress itself using various types of clothing. They demonstrated their technology by presenting a video that shows an animated figure gracefully putting on clothing, and the most interesting part is that it learned how […]

Read More