STACKQUADRANT

Fine-tuning Tools

Tools for fine-tuning, training, and adapting foundation models

52 repos

NVIDIA-NeMo/Curator

6.2

Scalable data pre processing and curation toolkit for LLMs

1.5k253Python

utkuozdemir/nvidia_gpu_exporter

5.9

Nvidia GPU exporter for prometheus using nvidia-smi binary

1.5k146Go

SakanaAI/text-to-lora

5.0

Hypernetworks that adapt LLMs for specific benchmark tasks using only textual task description as the input

1.3k87Python

datadreamer-dev/DataDreamer

4.9

DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models.   🤖💤

1.1k60Python

ARahim3/mlx-tune

5.5

Bringing the Unsloth experience to Mac users via Apple's MLX framework

1.1k67Python

amitshekhariitbhu/ai-engineering-interview-questions

5.6

Your Cheat Sheet for AI Engineering Interview – Questions and Answers.

1.0k191Markdown

InternScience/GraphGen

6.2

GraphGen: Enhancing Supervised Fine-Tuning for LLMs with Knowledge-Driven Synthetic Data Generation

1.0k78Python

volcengine/veScale

5.5

Byted PyTorch Distributed for Hyperscale Training of LLMs and RLs

1.0k61Python

louisfb01/start-llms

5.5

A complete guide to start and improve your LLM skills in 2026 with little background in the field and stay up-to-date with the latest news and state-of-the-art techniques!

960124

theredsix/cerebellum

5.0

Browser automation system that uses AI-driven planning to navigate web pages and perform goals.

85657Python

sail-sg/Adan

4.5

Adan: Adaptive Nesterov Momentum Algorithm for Faster Optimizing Deep Models

81371Python

IntelLabs/RAG-FiT

5.4

Framework for enhancing LLMs for RAG tasks using fine-tuning.

76961Python

ghimiresunil/LLM-PowerHouse-A-Curated-Guide-for-Large-Language-Models-with-Custom-Training-and-Inferencing

5.3

LLM-PowerHouse: Unleash LLMs' potential through curated tutorials, best practices, and ready-to-use code for custom training and inferencing.

727121Jupyter Notebook

tingaicompass/AI-Compass

5.3

“AI-Compass”将为社区指引在 AI 技术海洋中航行的方向,无论你是初学者还是进阶开发者,都能在这里找到通往 AI 各大方向的路径。旨在帮助开发者系统性地了解 AI 的核心概念、主流技术、前沿趋势,并通过实践掌握从理论到落地的全过程。

66984Python

feifeibear/long-context-attention

5.7

USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference

66479Python

yinizhilian/ICLR2025-Papers-with-Code

3.3

历年ICLR论文和开源项目合集,包含ICLR2021、ICLR2022、ICLR2023、ICLR2024、ICLR2025.

57029

rohan-paul/LLM-FineTuning-Large-Language-Models

3.6

LLM (Large Language Model) FineTuning

569137Jupyter Notebook

anarchy-ai/LLM-VM

4.8

irresponsible innovation. Try now at https://chat.dev/

491135Python

mallorbc/Finetune_LLMs

3.8

Repo for fine-tuning Casual LLMs

46086Python

FlagAI-Open/Aquila2

3.6

The official repo of Aquila2 series proposed by BAAI, including pretrained & chat large language models.

44531Python

InternLM/InternEvo

5.0

InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.

42068Python

awslabs/awsome-distributed-training

5.7

Collection of best practices, reference architectures, model training examples and utilities to train large models on AWS.

407182Shell

tigerlab-ai/tiger

4.3

Open Source LLM toolkit to build trustworthy LLM applications. TigerArmor (AI safety), TigerRAG (embedding, RAG), TigerTune (fine-tuning)

40127Jupyter Notebook

zhuhanqing/APOLLO

4.4

APOLLO: SGD-like Memory, AdamW-level Performance; MLSys'25 Oustanding Paper Honorable Mention

34418Python