Full text tutorial (requires MLExpert Pro): https://www.mlexpert.io/bootcamp/fine...
Getting bad predictions from your Tiny LLM? Learn how to finetune a small LLM (e.g. Phi2, TinyLlama) and (possibly) increase your model's performance. You'll understand how to set up a dataset, model, tokenizer, and LoRA adapter. We'll train the model (Tiny Llama) on a single GPU with custom data and evaluate the predictions.
AI Bootcamp (in preview): https://www.mlexpert.io/membership
Discord: / discord
Subscribe: http://bit.ly/venelinsubscribe
GitHub repository: https://github.com/curiousily/GetThi...
00:00 Intro
00:36 Text tutorial on MLExpert
01:01 Why finetuning Tiny LLM?
04:38 Prepare the dataset
09:46 Model & tokenizer setup
11:32 Token counts
12:41 Finetuning with LoRA
22:13 Training results & saving the model
24:00 Inference with the trained model
28:05 Evaluation
30:46 Conclusion
Join this channel to get access to the perks and support my work:
/ @venelin_valkov
#artificialintelligence #sentimentanalysis #llm #llama2 #chatgpt #gpt4 #python #chatbot