Reverse Dependencies of trl
The following projects have a declared dependency on trl:
- alignment-handbook — The Alignment Handbook
- arcee-align — The open source toolkit for finetuning and deploying LLMs
- argilla-v1 — Open-source tool for exploring, labeling, and monitoring data for NLP projects.
- atomgpt — atomgpt
- autotrain-advanced — no summary
- axolotl — LLM Trainer
- code-rl — Code RL
- competitions — Hugging Face Competitions
- dallm — Domain Adapted Language Model
- danoliterate — Benchmark of Generative Large Language Models in Danish
- datadreamer.dev — Prompt. Generate Synthetic Data. Train & Align Models.
- dbgpt-hub — DB-GPT-Hub: Text-to-SQL parsing with LLMs
- extralit — Open-source tool for accurate & fast scientific literature data extraction with LLM and human-in-the-loop.
- finetuna — no summary
- FineTune-Mistral — A package for fine-tuning Mistral model and generating responses.
- FineTune-Uunsloth-Mistral-7b — A package for fine-tuning Mistral model and generating responses.
- fms-acceleration-aadp — FMS Acceleration Plugin for Attention and Distributed Packing Optimizations
- fms-hf-tuning — FMS HF Tuning
- forgetnet — A package for applying differential privacy to model training using gradient shuffling and membership inference attack detection.
- ftraining — A fast LLM training module with QLoRA and GPU support.
- geniusrise-audio — audio bolts for geniusrise
- geniusrise-text — Text bolts for geniusrise
- geniusrise-vision — Huggingface bolts for geniusrise
- gguf-quantization — A package for quantizing LLMs
- glayout — A human language to analog layout API with support for different technologies.
- glmtuner — Fine-tuning ChatGLM-6B with PEFT
- huggify-data — This is a helper library to push data to HuggingFace.
- instructlab — Core package for interacting with InstructLab
- instructlab-training — Training Library
- lazyllm-llamafactory — Easy-to-use LLM fine-tuning framework
- liger-kernel — Efficient Triton kernels for LLM Training
- llama-trainer — Llama trainer utility
- llamafactory — Easy-to-use LLM fine-tuning framework
- llamafactory-songlab — Easy-to-use LLM fine-tuning framework
- llamagym — Fine-tune LLM agents with online reinforcement learning
- llm-finetuner — A package for fine-tuning LLMs using LoRA
- llm-toolkit — LLM Finetuning resource hub + toolkit
- llm-vm — An Open-Source AGI Server for Open-Source LLMs
- llmcompressor — A library for compressing large language models utilizing the latest techniques and research in the field for both training aware and post training techniques. The library is designed to be flexible and easy to use on top of PyTorch and HuggingFace Transformers, allowing for quick experimentation.
- llmcompressor-nightly — A library for compressing large language models utilizing the latest techniques and research in the field for both training aware and post training techniques. The library is designed to be flexible and easy to use on top of PyTorch and HuggingFace Transformers, allowing for quick experimentation.
- llmtuner — Easy-to-use LLM fine-tuning framework
- lm-buddy — Ray-centric library for finetuning and evaluation of (large) language models.
- mmlm — no summary
- ms-swift — Swift: Scalable lightWeight Infrastructure for Fine-Tuning
- nixietune — A semantic search embedding model fine-tuning tool
- optimum-neuron — Optimum Neuron is the interface between the Hugging Face Transformers and Diffusers libraries and AWS Trainium and Inferentia accelerators. It provides a set of tools enabling easy model loading, training and inference on single and multiple neuron core settings for different downstream tasks.
- otter-ai — Otter: A Multi-Modal Model with In-Context Instruction Tuning
- predacons — A python library based on transformers for transfer learning
- process-supervision-torch — Process SuperVision - Pytorch
- pykoi — pykoi: Active learning in one unified interface
- qlora-tunner — A Python package for Qlora fine-tuning open-source Large language models
- quasarx — quasar - Pytorch
- query-package-documentation — A package to explore documentations
- questionnaire-mistral — no summary
- rewardbench — Tools for evaluating reward models
- sft-dpo-qlora — SFT-DPO-QLora Trainer Package
- simplifine-alpha — An easy to use, open-source LLM finetuning library that handles all the complexities of the process for you.
- social-llama — Social Llama
- superduper-transformers — Transformers is a popular AI framework, and we have incorporated native support for Transformers to provide essential Large Language Model (LLM) capabilities.
- symbolicai — A Neuro-Symbolic Framework for Large Language Models
- text2text — Text2Text Language Modeling Toolkit
- TokenProbs — Extract token-level probabilities from LLMs for classification-type outputs.
- translation-canvas — Translation Canvas - A tool for evaluating and visualizing machine translation models
- tuningtron — Library for fine-tuning large language models
- UnBIAS — A package based on LLMs for detecting bias, performing named entity, and debiasing text.
- unsloth — 2-5X faster LLM finetuning
- unsloth-zoo — Utils for Unsloth
- vietlm — An experiment for large models specific to Vietnamese, including language models and more.
- yival — YiVal is an open-source project designed to revolutionize the way developers and researchers evaluate and refine AI models.
- zarth-utils — Package used for my personal development on ML projects.
1