OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.
-
Updated
Feb 29, 2024 - Python
OpenAssistant is a chat-based assistant that understands tasks, can interact with third-party systems, and retrieve information dynamically to do so.
Unify Efficient Fine-tuning of 100+ LLMs
The official GitHub page for the survey paper "A Survey of Large Language Models".
Official release of InternLM2 7B and 20B base and chat models. 200K context support
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
Robust recipes to align language models with human and AI preferences
Argilla is a collaboration platform for AI engineers and domain experts that require high-quality outputs, full data ownership, and overall efficiency.
A Doctor for your data
A curated list of reinforcement learning with human feedback resources (continually updated)
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
[NeurIPS 2023] ImageReward: Learning and Evaluating Human Preferences for Text-to-image Generation
Xtreme1 is an all-in-one data labeling and annotation platform for multimodal data training and supports 3D LiDAR point cloud, image, and LLM.
Aligning Large Language Models with Human: A Survey
Implementation of ChatGPT RLHF (Reinforcement Learning with Human Feedback) on any generation model in huggingface's transformer (blommz-176B/bloom/gpt/bart/T5/MetaICL)
聚宝盆(Cornucopia): 中文金融系列开源可商用大模型,并提供一套高效的金融垂直领域LLM训练框架(Pretraining、SFT、RLHF、Quantize等)
Add a description, image, and links to the rlhf topic page so that developers can more easily learn about it.
To associate your repository with the rlhf topic, visit your repo's landing page and select "manage topics."