Home
Information

AI Dataset Collection

Large-scale datasets and benchmarks for training, evaluating, and testing models to measure

Tools

Intelligent Document Recognition

Comprehensive Text Extraction and Document Processing Solutions for Users

AI Tutorial

safe-rlhf

Public

Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback

Creat2023-05-15T19:47:08
Update2025-03-27T11:11:18
https://pku-beaver.github.io
1.5K
Stars
3
Stars Increase

Related projects