This is handwritten Notes from first session( READ_2)
What I learned :
Attention
Encoder & Decoder Model ( architecture)
BERT Model
Context Length
Lecture_transformer_02.pdf