Myvideo

Guest

Login

Attention in transformers, visually explained | Chapter 6, Deep Learning

Uploaded By: Myvideo
1 view
0
0 votes
0

Demystifying attention, the key mechanism inside transformers and LLMs. Instead of sponsored ad reads, these lessons are funded directly by viewers: Special thanks to these supporters: #thanks An equally valuable form of support is to simply share the videos. Demystifying self-attention, multiple heads, and cross-attention. Instead of sponsored ad reads, these lessons are funded directly by viewers: The first pass for the translated subtitles here is machine-generated, and therefore notably imperfect. To contribute edits or fixes, visit And yes, at 22:00 (and elsewhere), “breaks“ is a typo. ------------------ Here are a few other relevant resources Build a GPT from scratch, by Andrej Karpathy If you want a conceptual understanding of language models from the ground up, @vcubingx jus

Share with your friends

Link:

Embed:

Video Size:

Custom size:

x

Add to Playlist:

Favorites
My Playlist
Watch Later