Mathematics Behind Large Language Models and Transformers

Deep Dive into Transformer Mathematics: From Tokenization to Multi-Head Attention to Masked Language Modeling & Beyond
4.40 (373 reviews)
Udemy
platform
English
language
Other
category
instructor
Mathematics Behind Large Language Models and Transformers
2,014
students
4.5 hours
content
Jun 2024
last update
$79.99
regular price

What you will learn

Mathematics Behind Large Language Models

Positional Encodings

Multi Head Attention

Query, Value and Key Matrix

Attention Masks

Masked Language Modeling

Dot Products and Vector Alignments

Nature of Sine and Cosine functions in Positional Encodings

How models like ChatGPT work under the hood

Bidirectional Models

Context aware word representations

Word Embeddings

How dot products work

Matrix multiplication

Programatically Create tokens

Screenshots

Mathematics Behind Large Language Models and Transformers - Screenshot_01Mathematics Behind Large Language Models and Transformers - Screenshot_02Mathematics Behind Large Language Models and Transformers - Screenshot_03Mathematics Behind Large Language Models and Transformers - Screenshot_04
6029496
udemy ID
6/18/2024
course created date
7/15/2024
course indexed date
Bot
course submited by
Mathematics Behind Large Language Models and Transformers - | Comidoc