LaMini-LM - Mini Models Maxi Data!
Colab LaMini-LM Neo 1.3B: https://colab.research.google.com/drive/1JkbeqGDp_UIi12lv0CltJhFyzIqvY_1N?usp=sharing
Colab LaMini-LM GPT1.5B: https://colab.research.google.com/drive/1wAfYXYECNIeub0hS05c7BRTY2a26qd53?usp=sharing
Colab LaMini-Flan-T5-783M: https://colab.research.google.com/drive/1fJrwbqYFQa1wJ3xJelZ9gjTxcDnCqumb?usp=sharing
Github: https://github.com/mbzuai-nlp/LaMini-LM
Dataset: https://huggingface.co/datasets/MBZUAI/LaMini-instruction/
Paper: https://arxiv.org/abs/2304.14402
In this video I go through the paper LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions and examine how they created the dataset and models for this paper and project.
00:00 Intro
00:45 Key Idea
01:23 Diagram
01:45 Dataset
02:10 Hugging Face Dataset
02:27 Trained on a lot of Models
03:05 Paper
04:36 Prompts on ChatGPT
09:37 Code Time