Skip to content

Commit 61b4d3c

Browse files
authored
Update index.md
1 parent becb683 commit 61b4d3c

File tree

1 file changed

+2
-2
lines changed
  • content/posts/2025-01-22-decision-tree-reward-model

1 file changed

+2
-2
lines changed

content/posts/2025-01-22-decision-tree-reward-model/index.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -237,8 +237,8 @@ Looking at the decision trees fitted to our multi-objective reward models:
237237
* The Llama-3.1-8B tree, while also using helpfulness at its root, incorporates correctness, verbosity, and complexity in its deeper nodes. This mixed-attribute approach suggests the model has learned
238238

239239

240-
## Evaluation on RewardBench
241-
We evaluate our decision-tree-based reward models on [Reward-Bench](https://huggingface.co/spaces/allenai/reward-bench), a comprehensive benchmark designed to assess reward model performance across multiple dimensions of LLM alignment. Reward-Bench evaluates models on four key aspects: general chat quality, challenging chat scenarios, safety considerations, and reasoning capabilities.
240+
## Evaluation on RewardBench v1 (Jan 2025)
241+
We evaluate our decision-tree-based reward models on [RewardBench v1](https://huggingface.co/spaces/allenai/reward-bench) (Note: the leaderboard was updated to v2 in June 2025), a comprehensive benchmark designed to assess reward model performance across multiple dimensions of LLM alignment. Reward-Bench evaluates models on four key aspects: general chat quality, challenging chat scenarios, safety considerations, and reasoning capabilities.
242242

243243

244244
Rank | Model | Base Model | Method | Overall Score | Chat | Chat Hard | Safety | Reasoning |

0 commit comments

Comments
 (0)