In this lecture, we build the model architecture for the classification fine-tuning project.
We modify the original LLM architecture by adding a classification head to the architecture design.
We also load OpenAI GPT-2 pretrained weights into our modified model architecture.
0:00 Recap of classification fine-tuning so far
6:06 Loading OpenAI GPT-2 pretrained weights
16:13 Adding classification head to model architecture
20:07 Select layers which want to fine-tune
25:01 Coding the finetuning architecture
31:44 Extracting last token output
33:12 Next steps
As we learn AI/ML/DL the material, we will share thoughts on what is actually useful in industry and what has become irrelevant. We will also share a lot of information on which subject contains open areas of research. Interested students can also start their research journey there.
Students who are confused or stuck in their ML journey, maybe courses and offline videos are not inspiring enough. What might inspire you is if you see someone else learning and implementing machine learning from scratch.
No cost. No hidden charges. Pure old school teaching and learning.