FedRLHF: A Convergence-Guaranteed Federated Framework for Privacy-Preserving and Personalized RLHF

Read the full article See related articles

Listed in

This article is not in any list yet, why not save it to one of your lists.
Log in to save this article

Abstract

In the era of increasing privacy concerns and demand for personalized experiences, traditional Reinforcement Learning with Human Feedback (RLHF) frameworks face significant challenges due to their reliance on centralized data. We introduce Federated Reinforcement Learning with Human Feedback (FedRLHF), a novel framework that decentralizes the RLHF process.FedRLHF enables collaborative policy learning across multiple clients, such as Large Language Models (LLMs) finetuning, without sharing raw data or human feedback, thereby ensuring robust privacy preservation.Leveraging federated reinforcement learning, each client integrates human feedback locally into reward functions and updates their policies through personalized RLHF processes. We establish rigorous theoretical foundations for FedRLHF, providing convergence guarantees, and deriving sample complexity bounds that scale efficiently with the number of clients. Empirical evaluations on the MovieLens and IMDb datasets demonstrate that FedRLHF preserves user privacy, achieves performance on par with centralized RLHF, and enhances personalization across diverse client environments.

Article activity feed