
amazon-sagemaker-llm-fine-tuning-remote-decorator
None
Stars: 57

README:
Important: The scope of these notebook examples is to showcase interactive experience with SageMaker AI capabilities and @remote decorator, for Small/Large Language Models fine-tuning by using different distribution techniques, such as FSDP, and DDP.
In this example we will go through the steps required for interactively fine-tuning foundation models on Amazon SageMaker AI by using @remote decorator for executing Training jobs.
You can run this repository from Amazon SageMaker Studio or from your local IDE.
For additional information, take a look at the AWS Blog Fine-tune Falcon 7B and other LLMs on Amazon SageMaker with @remote decorator
The notebooks are currently using the latest PyTorch Training Container available for the region us-east-1
. If you are running the notebooks in a different region, make sure to update the ImageUri in the file config.yaml.
Python version used in the training container: Python 3.11
- Navigate [Available Deep Learning Containers Images](Available Deep Learning Containers Images)
- Select the right Hugging Face TGI container for model training based on your selected region
- Update ImageUri in the file config.yaml
- [Supervised - QLoRA] Falcon-7B
- [Supervised - QLoRA, FSDP] Llama-13B
- [Self-supervised - QLoRA, FSDP] Llama-13B
- [Self-supervised - QLoRA] Mistral-7B
- [Supervised - QLoRA, FSDP] Mixtral-8x7B
- [Supervised - QLoRA, DDP] Code-Llama 13B
- [Supervised - QLORA, DDP] Llama-3 8B
- [Supervised - QLoRA, DDP] Llama-3.1 8B
- [Supervised - QLoRA, DDP] Arcee AI Llama-3.1 Supernova Lite
- [Supervised - QLoRA] Llama-3.2 1B
- [Supervised - QLoRA] Llama-3.2 3B
- [Supervised - QLoRA, FSDP] Codestral-22B
- [Supervised - LoRA] TinyLlama 1.1B
- [Supervised - LoRA] Arcee Lite 1.5B
- [Supervised - LoRA] SmolLM2-1.7B-Instruct
- [Supervised - QLORA, FSDP] Qwen 2.5 7B
- [Supervised - QLORA] Falcon3 3B
- [Supervised - QLORA, FSDP] Falcon3 7B
- [Supervised - QLORA, FSDP] Llama-3.1 70B
- [Self-supervised - DoRA, FSDP] Mistral-7B v0.3
- [Supervised - QLORA, FSDP] Llama-3.3 70B
- [Supervised - QLORA, FSDP] OpenCoder-8B-Instruct
- [Supervised - QLORA, FSDP] DeepSeek-R1-Distill-Qwen-32B
- [Supervised - QLORA, FSDP] DeepSeek-R1-Distill-Llama-70B
- [Supervised - QLORA, FSDP] DeepSeek-R1-Distill-Llama-8B
- [Supervised - QLORA, DDP] DeepSeek-R1-Distill-Qwen-1.5B
- [Supervised - QLORA, FSDP] DeepSeek-R1-Distill-Qwen-7B
- [Supervised - QLORA, FSDP] Mistral-Small-24B-Instruct-2501
return cloudpickle.loads(bytes_to_deserialize)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
Traceback (most recent call last): in deserialize return cloudpickle.loads(bytes_to_deserialize)
YYYY-MM-DDThh:mm:ss
AttributeError: Can't get attribute '_function_setstate' on <module 'cloudpickle.cloudpickle' from '/opt/conda/lib/python3.11/site-packages/cloudpickle/cloudpickle.py'>
Align your cloudpickle
local version to the container one, by including in your requirements.txt
:
cloudpickle==x.x.x
Where x.x.x is the version you want to install.
For Tasks:
Click tags to check more tools for each tasksFor Jobs:
Alternative AI tools for amazon-sagemaker-llm-fine-tuning-remote-decorator
Similar Open Source Tools

awesome-deliberative-prompting
The 'awesome-deliberative-prompting' repository focuses on how to ask Large Language Models (LLMs) to produce reliable reasoning and make reason-responsive decisions through deliberative prompting. It includes success stories, prompting patterns and strategies, multi-agent deliberation, reflection and meta-cognition, text generation techniques, self-correction methods, reasoning analytics, limitations, failures, puzzles, datasets, tools, and other resources related to deliberative prompting. The repository provides a comprehensive overview of research, techniques, and tools for enhancing reasoning capabilities of LLMs.

core
OpenSumi is a framework designed to help users quickly build AI Native IDE products. It provides a set of tools and templates for creating Cloud IDEs, Desktop IDEs based on Electron, CodeBlitz web IDE Framework, Lite Web IDE on the Browser, and Mini-App liked IDE. The framework also offers documentation for users to refer to and a detailed guide on contributing to the project. OpenSumi encourages contributions from the community and provides a platform for users to report bugs, contribute code, or improve documentation. The project is licensed under the MIT license and contains third-party code under other open source licenses.

efficient-transformers
Efficient Transformers Library provides reimplemented blocks of Large Language Models (LLMs) to make models functional and highly performant on Qualcomm Cloud AI 100. It includes graph transformations, handling for under-flows and overflows, patcher modules, exporter module, sample applications, and unit test templates. The library supports seamless inference on pre-trained LLMs with documentation for model optimization and deployment. Contributions and suggestions are welcome, with a focus on testing changes for model support and common utilities.

Awesome-Text2SQL
Awesome Text2SQL is a curated repository containing tutorials and resources for Large Language Models, Text2SQL, Text2DSL, Text2API, Text2Vis, and more. It provides guidelines on converting natural language questions into structured SQL queries, with a focus on NL2SQL. The repository includes information on various models, datasets, evaluation metrics, fine-tuning methods, libraries, and practice projects related to Text2SQL. It serves as a comprehensive resource for individuals interested in working with Text2SQL and related technologies.

LLM-from-scratch
This repository contains notes on re-implementing some LLM models from scratch. It includes steps to pre-train a super mini LLaMA 3 model, implement LoRA from scratch using PyTorch, and work on implementing the 'generate' method.

IntelliQ
IntelliQ is an open-source project aimed at providing a multi-turn question-answering system based on a large language model (LLM). The system combines advanced intent recognition and slot filling technology to enhance the depth of understanding and accuracy of responses in conversation systems. It offers a flexible and efficient solution for developers to build and optimize various conversational applications. The system features multi-turn dialogue management, intent recognition, slot filling, interface slot technology for real-time data retrieval and processing, adaptive learning for improving response accuracy and speed, and easy integration with detailed API documentation supporting multiple programming languages and platforms.

gorilla
Gorilla is a tool that enables LLMs to use tools by invoking APIs. Given a natural language query, Gorilla comes up with the semantically- and syntactically- correct API to invoke. With Gorilla, you can use LLMs to invoke 1,600+ (and growing) API calls accurately while reducing hallucination. Gorilla also releases APIBench, the largest collection of APIs, curated and easy to be trained on!

svelte-commerce
Svelte Commerce is an open-source frontend for eCommerce, utilizing a PWA and headless approach with a modern JS stack. It supports integration with various eCommerce backends like MedusaJS, Woocommerce, Bigcommerce, and Shopify. The API flexibility allows seamless connection with third-party tools such as payment gateways, POS systems, and AI services. Svelte Commerce offers essential eCommerce features, is both SSR and SPA, superfast, and free to download and modify. Users can easily deploy it on Netlify or Vercel with zero configuration. The tool provides features like headless commerce, authentication, cart & checkout, TailwindCSS styling, server-side rendering, proxy + API integration, animations, lazy loading, search functionality, faceted filters, and more.

CuMo
CuMo is a project focused on scaling multimodal Large Language Models (LLMs) with Co-Upcycled Mixture-of-Experts. It introduces CuMo, which incorporates Co-upcycled Top-K sparsely-gated Mixture-of-experts blocks into the vision encoder and the MLP connector, enhancing the capabilities of multimodal LLMs. The project adopts a three-stage training approach with auxiliary losses to stabilize the training process and maintain a balanced loading of experts. CuMo achieves comparable performance to other state-of-the-art multimodal LLMs on various Visual Question Answering (VQA) and visual-instruction-following benchmarks.

Recommendation-Systems-without-Explicit-ID-Features-A-Literature-Review
This repository is a collection of papers and resources related to recommendation systems, focusing on foundation models, transferable recommender systems, large language models, and multimodal recommender systems. It explores questions such as the necessity of ID embeddings, the shift from matching to generating paradigms, and the future of multimodal recommender systems. The papers cover various aspects of recommendation systems, including pretraining, user representation, dataset benchmarks, and evaluation methods. The repository aims to provide insights and advancements in the field of recommendation systems through literature reviews, surveys, and empirical studies.

suaveui
SuaveUI is an experimental Progressive Web App chat user interface designed for interacting with local AI models. It provides a platform for users to easily communicate with AI models in a chat-like environment. The tool is built using React for the user interface and Node.js for the backend. Users can run SuaveUI using Docker or by cloning the repository and running a server. The project is still in the early alpha stage and is being actively developed to enhance its functionality and features.

IvyGPT
IvyGPT is a medical large language model that aims to generate the most realistic doctor consultation effects. It has been fine-tuned on high-quality medical Q&A data and trained using human feedback reinforcement learning. The project features full-process training on medical Q&A LLM, multiple fine-tuning methods support, efficient dataset creation tools, and a dataset of over 300,000 high-quality doctor-patient dialogues for training.

ai-to-pptx
Ai-to-pptx is a tool that uses AI technology to automatically generate PPTX, and supports online editing and exporting of PPTX. Main functions: - 1 Use large language models such as ChatGPT to generate outlines - 2 The generated content allows users to modify again - 3 Different templates can be selected when generating PPTX - 4 Support online editing of PPTX text content, style, pictures, etc. - 5 Supports exporting PPTX, PDF, PNG and other formats - 6 Support users to set their own LOGO and related background pictures to create their own exclusive PPTX style - 7 Support users to design their own templates and upload them to the sharing platform for others to use

FuseAI
FuseAI is a repository that focuses on knowledge fusion of large language models. It includes FuseChat, a state-of-the-art 7B LLM on MT-Bench, and FuseLLM, which surpasses Llama-2-7B by fusing three open-source foundation LLMs. The repository provides tech reports, releases, and datasets for FuseChat and FuseLLM, showcasing their performance and advancements in the field of chat models and large language models.