Week 3 - LLM - PreTraining
Week 3 - LLM - PreTraining
Fine Tuning
is like a student learning to use language to
perform complex tasks in high school and college.
In-Context Learning
is like a working professional trying to
figure out his/her manager’s instructions
Zero Shot vs Few Shot
In-Context Learning (few shot learning)
The LLM Landscape
BERT GPT3 Llama 2
Year 2018 2020 2023
Developer Google OpenAI Meta
Parameters 110 M, 340 M 175 B 7 B, 13 B, 70 B
Architecture Encoder only Decoder only Decoder only
Embedding Size 768 12888 3204
Context Length 512 2048 4000
Tokenization WordPiece BPE SentencePiece
Use Case Classification, NER, Q&A Text Generation Text Generation
The GPT Models
[RLHF]
NON-GRADED TASKS
Write a 400-word blog on the BERT model
that a non-CS person can understand
Teach the concept of word embeddings
and sentence similarity to at least 3 first year students
- Find their PCA components (n=2) and see if they form any clusters.
- Make random changes in the model parameters and see its effect