A curated list of reinforcement learning with human feedback resources[awesome-RLHF-Turkish] (continually updated)
-
Updated
Apr 27, 2023
A curated list of reinforcement learning with human feedback resources[awesome-RLHF-Turkish] (continually updated)
[ICML 2024] Code for the paper "Confronting Reward Overoptimization for Diffusion Models: A Perspective of Inductive and Primacy Biases"
Reinforcement Learning from Human Feedback with 🤗 TRL
Aligning LLM Agents by Learning Latent Preference from User Edits
[ NeurIPS 2023 ] Official Codebase for "Aligning Synthetic Medical Images with Clinical Knowledge using Human Feedback"
The Prism Alignment Project
BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).
[CVPR 2024] Code for the paper "Using Human Feedback to Fine-tune Diffusion Models without Any Reward Model"
Product analytics for AI Assistants
Let's build better datasets, together!
Implementation of Reinforcement Learning from Human Feedback (RLHF)
The ParroT framework to enhance and regulate the Translation Abilities during Chat based on open-sourced LLMs (e.g., LLaMA-7b, Bloomz-7b1-mt) and human written translation and evaluation data.
Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.
A curated list of reinforcement learning with human feedback resources (continually updated)
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
Add a description, image, and links to the human-feedback topic page so that developers can more easily learn about it.
To associate your repository with the human-feedback topic, visit your repo's landing page and select "manage topics."