HFEPX Hub
CS.CR Human Feedback And Eval Papers
Updated from current HFEPX corpus (Mar 8, 2026). 57 papers are grouped in this hub page.
Read Full Context
Updated from current HFEPX corpus (Mar 8, 2026). 57 papers are grouped in this hub page. Common evaluation modes: Automatic Metrics, Llm As Judge. Most common rater population: Domain Experts. Common annotation unit: Trajectory. Frequent quality control: Calibration. Frequently cited benchmark: AdvBench. Common metric signal: jailbreak success rate. Use this page to compare protocol setup, judge behavior, and labeling design decisions before running new eval experiments. Newest paper in this set is from Feb 27, 2026.