Multimodal Few-Shot Learning with Frozen Language Models | Paper Explained
❤️ Become The AI Epiphany Patreon ❤️ ►
In this video I cover “Multimodal Few-Shot Learning with Frozen Language Models“ from DeepMind. They introduce Frozen - which is able to handle both visual and textual inputs and shows good generalization capabilities to novel visual question answering datasets combined with fast binding mechanisms even though it was only trained on image captioning.
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
✅ Paper:
▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬▬
⌚️ Timetable:
00:00 Intro
02:20 GPT-3 and emerging few-shot properties
04:20 Training procedure for Frozen
3 views
46
14
3 years ago 00:19:59 1
What Can Huge Neural Networks do?
3 years ago 00:56:52 5
Deep Learning, Transformers, and the Consequences of Scale with Oriol Vinyals - #546
4 years ago 00:17:05 1
AI Weekly Update Overview - July 15th, 2021
4 years ago 00:13:11 1
Multimodal Few-Shot Learning with Frozen Language Models
4 years ago 00:21:01 3
Multimodal Few-Shot Learning with Frozen Language Models | Paper Explained