LLM-FineTuning-Large-Language-Models

LLM-FineTuning-Large-Language-Models

LLM (Large Language Model) FineTuning

Stars: 319

Visit
 screenshot

This repository contains projects and notes on common practical techniques for fine-tuning Large Language Models (LLMs). It includes fine-tuning LLM notebooks, Colab links, LLM techniques and utils, and other smaller language models. The repository also provides links to YouTube videos explaining the concepts and techniques discussed in the notebooks.

README:

LLM (Large Language Models) FineTuning Projects and notes on common practical techniques

Find me here..


Fine-tuning LLM (and YouTube Video Explanations)

Notebook ๐ŸŸ  YouTube Video
Llama-3 Finetuning on custom dataset with unsloth Youtube Link
CodeLLaMA-34B - Conversational Agent Youtube Link
Inference Yarn-Llama-2-13b-128k with KV Cache to answer quiz on very long textbook Youtube Link
Mistral 7B FineTuning with_PEFT and QLORA Youtube Link
Falcon finetuning on openassistant-guanaco Youtube Link
Fine Tuning Phi 1_5 with PEFT and QLoRA Youtube Link
Web scraping with Large Language Models (LLM)-AnthropicAI + LangChainAI Youtube Link

Fine-tuning LLM

Notebook Colab
๐Ÿ“Œ Gemma_2b_finetuning_ORPO_full_precision Open In Colab
๐Ÿ“Œ Jamba_Finetuning_Colab-Pro Open In Colab
๐Ÿ“Œ Finetune codellama-34B with QLoRA Open In Colab
๐Ÿ“Œ Mixtral Chatbot with Gradio
๐Ÿ“Œ togetherai api to run Mixtral Open In Colab
๐Ÿ“Œ Integrating TogetherAI with LangChain ๐Ÿฆ™ Open In Colab
๐Ÿ“Œ Mistral-7B-Instruct_GPTQ - Finetune on finance-alpaca dataset ๐Ÿฆ™ Open In Colab
๐Ÿ“Œ Mistral 7b FineTuning with DPO Direct_Preference_Optimization Open In Colab
๐Ÿ“Œ Finetune llama_2_GPTQ
๐Ÿ“Œ TinyLlama with Unsloth and_RoPE_Scaling dolly-15 dataset Open In Colab
๐Ÿ“Œ Tinyllama fine-tuning with Taylor_Swift Song lyrics Open In Colab

LLM Techniques and utils - Explained

LLM Concepts
๐Ÿ“Œ DPO (Direct Preference Optimization) training and its datasets
๐Ÿ“Œ 4-bit LLM Quantization with GPTQ
๐Ÿ“Œ Quantize with HF Transformers
๐Ÿ“Œ Understanding rank r in LoRA and related Matrix_Math
๐Ÿ“Œ Rotary Embeddings (RopE) is one of the Fundamental Building Blocks of LlaMA-2 Implementation
๐Ÿ“Œ Chat Templates in HuggingFace
๐Ÿ“Œ How is Mixtral 8x7B is a dense 47Bn param model
๐Ÿ“Œ The concept of validation log perplexity in LLM training - a note on fundamentals.
๐Ÿ“Œ Why we need to identify target_layers for LoRA/QLoRA
๐Ÿ“Œ Evaluate Token per sec
๐Ÿ“Œ traversing through nested attributes (or sub-modules) of a PyTorch module
๐Ÿ“Œ Implementation of Sparse Mixtures-of-Experts layer in PyTorch from Mistral Official Repo
๐Ÿ“Œ Util method to extract a specific token's representation from the last hidden states of a transformer model.
๐Ÿ“Œ Convert PyTorch model's parameters and tensors to half-precision floating-point format
๐Ÿ“Œ Quantizing ๐Ÿค— Transformers models with the GPTQ method
๐Ÿ“Œ Quantize Mixtral-8x7B so it can run in 24GB GPU
๐Ÿ“Œ What is GGML or GGUF in the world of Large Language Models ?

Other Smaller Language Models

For Tasks:

Click tags to check more tools for each tasks

For Jobs:

Alternative AI tools for LLM-FineTuning-Large-Language-Models

Similar Open Source Tools

For similar tasks

For similar jobs