Myvideo

Guest

Login

AI 360: 01/03/2021. Unified Transformer, Sebastian Ruder, OpenAI's DALL-E, GLOM and StudioGAN

Uploaded By: Myvideo
10 views
0
0 votes
0

For the full experience, and links to everything referenced, visit our website: FAIR proposes Unified Transformer Recently, we have seen that Transformers have led to a paradigm shift in AI and NLP research (now even computer vision). Multi-modal research has recently employed Transformers in large Vision/Language Pretraining frameworks (such as VILBERT, VLP etc). Models such as these are usually only trained on one or two pre-training tasks. Facebook AI Research (FAIR) propose a multi-modal model they call the Unified Transformer (UniT), which is a Transformer based model jointly trained on 7 different tasks: object detection, VQA, SNLI-VE, MNLI, QNLI, QQP and SST-2. The architecture, which achieves comparable results to task specific Transformer based models with a signficantly reduced parameter set uses two Transformer encoders and one Transformer decoder. At a very high level, one Transformer encoder is responsible for encoding the image, and the other for encoding the text. T

Share with your friends

Link:

Embed:

Video Size:

Custom size:

x

Add to Playlist:

Favorites
My Playlist
Watch Later