TRL

Open Source Free Trial

Hugging Face library for training language models with RLHF, SFT, and DPO

TRL (Transformer Reinforcement Learning) is the standard Hugging Face library for fine-tuning language models. It supports supervised fine-tuning (SFT), reinforcement learning from human feedback (RLHF), direct preference optimization (DPO), and other alignment techniques. Built on top of Transformers and integrates with PEFT for parameter-efficient training.

Pricing: Free

HQ 🇺🇸 United States
Screenshot of TRL webpage

Is your product missing?

Add it here →