{"id":3093066,"date":"2024-02-02T06:45:04","date_gmt":"2024-02-02T11:45:04","guid":{"rendered":"https:\/\/wordpress-1016567-4521551.cloudwaysapps.com\/plato-data\/fine-tuning-a-tiny-llama-model-with-unsloth\/"},"modified":"2024-02-02T06:45:04","modified_gmt":"2024-02-02T11:45:04","slug":"fine-tuning-a-tiny-llama-model-with-unsloth","status":"publish","type":"station","link":"https:\/\/platodata.io\/plato-data\/fine-tuning-a-tiny-llama-model-with-unsloth\/","title":{"rendered":"Fine-tuning A Tiny-Llama Model with Unsloth"},"content":{"rendered":"

Introduction<\/h2>\n

After the Llama and Mistral models were released, the open-source LLMs took the limelight out of OpenAI. Since then, multiple models have been released based on Llama and Mistral architecture, performing on par with proprietary models like GPT-3.5 Turbo, Claude, Gemini, etc. However, these models are too large to be used in consumer hardware.<\/p>\n

But lately, there has been an emergence of a new class of LLMs. These are the LLMs in the sub-7B parameter category. Fewer parameters make them compact enough to be run in consumer hardware while keeping efficiency comparable to the 7B models. Models like Tiny-Llama-1B, Microsoft\u2019s Phi-2, and Alibaba\u2019s Qwen-3b can be great substitutes for larger models to run locally or deploy on edge. At the same time, fine-tuning is crucial to bring the best out of any base model for any downstream tasks.
Here, we will explore how to Fine-tune a base Tiny-Llama model<\/a> on a cleaned Alpaca dataset.<\/p>\n

\n
\"Fine-tuning<\/figure>\n<\/div>\n

Learning Objectives<\/h4>\n