Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
-
Updated
Jan 14, 2024 - Python
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
A curated list of reinforcement learning with human feedback resources (continually updated)
Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.
The ParroT framework to enhance and regulate the Translation Abilities during Chat based on open-sourced LLMs (e.g., LLaMA-7b, Bloomz-7b1-mt) and human written translation and evaluation data.
Implementation of Reinforcement Learning from Human Feedback (RLHF)
Let's build better datasets, together!
Product analytics for AI Assistants
[CVPR 2024] Code for the paper "Using Human Feedback to Fine-tune Diffusion Models without Any Reward Model"
BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).
The Prism Alignment Project
[ NeurIPS 2023 ] Official Codebase for "Aligning Synthetic Medical Images with Clinical Knowledge using Human Feedback"
Aligning LLM Agents by Learning Latent Preference from User Edits
Reinforcement Learning from Human Feedback with 🤗 TRL
A curated list of reinforcement learning with human feedback resources[awesome-RLHF-Turkish] (continually updated)
Add a description, image, and links to the human-feedback topic page so that developers can more easily learn about it.
To associate your repository with the human-feedback topic, visit your repo's landing page and select "manage topics."