HomeAI Tutorial
Information

AI Dataset Collection

Large-scale datasets and benchmarks for training, evaluating, and testing models to measure

Tools

Intelligent Document Recognition

Comprehensive Text Extraction and Document Processing Solutions for Users

Reinforcement-Learning-for-Human-Feedback-RLHF

Public

This repository contains the implementation of a Reinforcement Learning with Human Feedback (RLHF) system using custom datasets. The project utilizes the trlX library for training a preference model that integrates human feedback directly into the optimization of language models.

Creat2024-08-17T15:27:37
Update2025-02-15T21:02:32
5
Stars
0
Stars Increase

Related projects