Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
-
Updated
Feb 20, 2025 - Python
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.
The official GitHub page for the survey paper "A Survey of Large Language Models".
Official release of InternLM series (InternLM, InternLM2, InternLM2.5, InternLM3).
Robust recipes to align language models with human and AI preferences
Argilla is a collaboration tool for AI engineers and domain experts to build high-quality datasets
A curated list of reinforcement learning with human feedback resources (continually updated)
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
The easiest tool for fine-tuning LLM models, synthetic data generation, and collaborating on datasets.
A Doctor for your data
Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verified research papers.
Align Anything: Training All-modality Model with Feedback
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
[NeurIPS 2023] ImageReward: Learning and Evaluating Human Preferences for Text-to-image Generation
Add a description, image, and links to the rlhf topic page so that developers can more easily learn about it.
To associate your repository with the rlhf topic, visit your repo's landing page and select "manage topics."