wangclnlp commited on
Commit
84fe5e4
·
verified ·
1 Parent(s): f8bfb75

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +11 -3
README.md CHANGED
@@ -12,9 +12,9 @@ tags:
12
  ---
13
  # Introduction
14
 
15
- This repository contains the released models for the paper [GRAM: A Generative Foundation Reward Model for Reward Generalization 📝]().
16
 
17
- <img src="https://raw.githubusercontent.com/wangclnlp/GRAM/refs/heads/main/gram.png?token=GHSAT0AAAAAAC5DHKJKFOGQKURCJNSPUTJG2CRBUJQ" width="1000px"></img>
18
 
19
  This training process is introduced above. Traditionally, these models are trained using labeled data, which can limit their potential. In this study, we propose a new method that combines both labeled and unlabeled data for training reward models. We introduce a generative reward model that first learns from a large amount of unlabeled data and is then fine-tuned with supervised data. Additionally, we demonstrate that using label smoothing during training improves performance by optimizing a regularized ranking loss. This approach bridges generative and discriminative models, offering a new perspective on training reward models. Our model can be easily applied to various tasks without the need for extensive fine-tuning. This means that when aligning LLMs, there is no longer a need to train a reward model from scratch with large amounts of task-specific labeled data. Instead, **you can directly apply our reward model or adapt it to align your LLM based on our [code](https://github.com/wangclnlp/GRAM/tree/main)**.
20
 
@@ -96,5 +96,13 @@ print({
96
 
97
  If you find this model helpful for your research, please cite GRAM:
98
  ```bash
99
- bib
 
 
 
 
 
 
 
 
100
  ```
 
12
  ---
13
  # Introduction
14
 
15
+ This repository contains the released models for the paper [GRAM: A Generative Foundation Reward Model for Reward Generalization 📝](https://arxiv.org/abs/2506.14175).
16
 
17
+ <img src="https://raw.githubusercontent.com/wangclnlp/GRAM/refs/heads/main/gram.png" width="1000px"></img>
18
 
19
  This training process is introduced above. Traditionally, these models are trained using labeled data, which can limit their potential. In this study, we propose a new method that combines both labeled and unlabeled data for training reward models. We introduce a generative reward model that first learns from a large amount of unlabeled data and is then fine-tuned with supervised data. Additionally, we demonstrate that using label smoothing during training improves performance by optimizing a regularized ranking loss. This approach bridges generative and discriminative models, offering a new perspective on training reward models. Our model can be easily applied to various tasks without the need for extensive fine-tuning. This means that when aligning LLMs, there is no longer a need to train a reward model from scratch with large amounts of task-specific labeled data. Instead, **you can directly apply our reward model or adapt it to align your LLM based on our [code](https://github.com/wangclnlp/GRAM/tree/main)**.
20
 
 
96
 
97
  If you find this model helpful for your research, please cite GRAM:
98
  ```bash
99
+ @misc{wang2025gram,
100
+ title={GRAM: A Generative Foundation Reward Model for Reward Generalization},
101
+ author={Chenglong Wang and Yang Gan and Yifu Huo and Yongyu Mu and Qiaozhi He and Murun Yang and Bei Li and Tong Xiao and Chunliang Zhang and Tongran Liu and Jingbo Zhu},
102
+ year={2025},
103
+ eprint={2506.14175},
104
+ archivePrefix={arXiv},
105
+ primaryClass={cs.CL},
106
+ url={https://arxiv.org/abs/2506.14175},
107
+ }
108
  ```