Skip to content

Latest commit

 

History

History
41 lines (31 loc) · 1.82 KB

README.md

File metadata and controls

41 lines (31 loc) · 1.82 KB

Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback

🤗 Preference Dataset | 📚 Documentation | 📄 Paper

This repository is the source code for the paper, Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback, where we introduce a routing framework that creates hybrid preferences with both LLM and human preference annotations to maximize performance on a given evaluation metric (e.g., RewardBench). We release this codebase to improve reproducibility of our work, and to aid researchers in constructing preference datasets in their research.

main_figure

Setup

Install the dependencies within your Python environment:

python -m venv venv
venv/bin/source activate
pip install -r requirements.txt

Documentation

Running the full pipeline involves several steps, some might need to be run on a TPU machine. Nevertheless, we wrote scripts to automate different parts of the pipeline. Please head over the docs directory for more information.

Citation

@article{miranda2024hybrid,
  title={{Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback}},
  author={Miranda, Lester James V and Wang, Yizhong and Elazar, Yanai and Kumar, Sachin and Pyatkin, Valentina and Brahman, Faeze and Smith, Noah A and Hajishirzi, Hannaneh and Dasigi, Pradeep},
  journal={arXiv preprint arXiv:2410.19133},
  year={2024}
}