Mathematics Behind Large Language Models and Transformers
Deep Dive into Transformer Mathematics: From Tokenization to Multi-Head Attention to Masked Language Modeling & Beyond
4.40 (373 reviews)

2,014
students
4.5 hours
content
Jun 2024
last update
$79.99
regular price
What you will learn
Mathematics Behind Large Language Models
Positional Encodings
Multi Head Attention
Query, Value and Key Matrix
Attention Masks
Masked Language Modeling
Dot Products and Vector Alignments
Nature of Sine and Cosine functions in Positional Encodings
How models like ChatGPT work under the hood
Bidirectional Models
Context aware word representations
Word Embeddings
How dot products work
Matrix multiplication
Programatically Create tokens
Screenshots




6029496
udemy ID
6/18/2024
course created date
7/15/2024
course indexed date
Bot
course submited by