Update README.md
Browse files
README.md
CHANGED
@@ -31,6 +31,9 @@ We evaluate GRM on the [reward model benchmark](https://huggingface.co/spaces/al
|
|
31 |
|
32 |
|
33 |
## Usage
|
|
|
|
|
|
|
34 |
```
|
35 |
import torch
|
36 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
@@ -56,7 +59,6 @@ with torch.no_grad():
|
|
56 |
reward = reward_tensor.cpu().detach().item()
|
57 |
```
|
58 |
|
59 |
-
**Note: loading llama3 model into 8 bit could lead to performance degradation.**
|
60 |
|
61 |
## Citation
|
62 |
If you find this model helpful for your research, please cite GRM
|
|
|
31 |
|
32 |
|
33 |
## Usage
|
34 |
+
**Note 1: Please download the `model.py` file from this repository to ensure the structure is loaded correctly and verify that the `v_head` is properly initialized.**
|
35 |
+
|
36 |
+
**Note 2: loading llama3 model into 8 bit could lead to performance degradation.**
|
37 |
```
|
38 |
import torch
|
39 |
from transformers import AutoTokenizer, AutoModelForSequenceClassification
|
|
|
59 |
reward = reward_tensor.cpu().detach().item()
|
60 |
```
|
61 |
|
|
|
62 |
|
63 |
## Citation
|
64 |
If you find this model helpful for your research, please cite GRM
|