Skip to content

Imran-ml/Fine-tuned-Llama-3.1

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 

Repository files navigation

Fine-Tuning LLaMA 3.1: Step-by-Step, Everything Explained in Detail

[Image Reference](https://hyperight.com/wp-content/uploads/2024/07/Meta-Llama-31-1024x576.webp)
Mental health is a critical aspect of overall well being for emotional, psychological and social dimensions.

We are going to use Unsloth because it significantly enhances the efficiency of fine-tuning large language models (LLMs) specially LLaMA and Mistral. With Unsloth, we can use advanced quantization techniques, such as 4-bit and 16-bit quantization, to reduce the memory and speed up both training and inference. This means we can deploy powerful models even on hardware with limited resources but without compromising on performance.

Additionally, Unsloth broad compatibility and customization options allow to do the quantization process to fit the specific needs of products. This flexibility combined with its ability to cut VRAM usage by up to 60%, makes Unsloth an essential tool in AI toolkit. Its not just about optimizing models its about making cutting-edge AI more accessible and efficient for real world applications.

For fine tuning, I used the following setup:
  • Torch 2.1.1 with CUDA 12.1 for efficient computation.
  • Unsloth to achieve 2X faster training speeds for the large language model (LLM).
  • H100 NVL GPU to handle the intensive processing requirement but you can use the less power GPU as well.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published