Multimodal Neurons in Artificial Neural Networks (w/ OpenAI Microscope, Research Paper Explained)
#openai #clip #microscope
OpenAI does a huge investigation into the inner workings of their recent CLIP model via faceted feature visualization and finds amazing things: Some neurons in the last layer respond to distinct concepts across multiple modalities, meaning they fire for photographs, drawings, and signs depicting the same concept, even when the images are vastly distinct. Through manual examination, they identify and investigate neurons corresponding to persons, geographical regions, religions, emotions, and much more. In this video, I go through the publication and then I present my own findings from digging around in the OpenAI Microscope.
OUTLINE:
0:00 - Intro & Overview
3:35 - OpenAI Microscope
7:10 - Categories of found neurons
11:10 - Person Neurons
13:00 - Donald Trump Neuron
17:15 - Emotion Neurons
22:45 - Region Neurons
26:40 - Sparse Mixture of Emotions
28:05 - Emotion Atlas
29:45 - Adversarial Typographic Attacks
31:55 - Stroop Test
33:10 - My Findings in OpenAI Microscope
33:30 - Superma