Add model card metadata and link to paper and code repository (#1)
Browse files- Add model card metadata and link to paper and code repository (ab146d1f70834483a042d8485d695be3d053cccc)
Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>
README.md
CHANGED
@@ -1,3 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
Quantization made by Richard Erkhov.
|
2 |
|
3 |
[Github](https://github.com/RichardErkhov)
|
@@ -6,14 +12,13 @@ Quantization made by Richard Erkhov.
|
|
6 |
|
7 |
[Request more models](https://github.com/RichardErkhov/quant_request)
|
8 |
|
|
|
|
|
9 |
|
10 |
SPPO-Llama-3-8B-Instruct-GPM-2B - bnb 8bits
|
11 |
- Model creator: https://huggingface.co/general-preference/
|
12 |
- Original model: https://huggingface.co/general-preference/SPPO-Llama-3-8B-Instruct-GPM-2B/
|
13 |
|
14 |
-
|
15 |
-
|
16 |
-
|
17 |
Original model description:
|
18 |
---
|
19 |
language:
|
@@ -68,7 +73,7 @@ model-index:
|
|
68 |
value: 8.01
|
69 |
name: exact match
|
70 |
source:
|
71 |
-
url: https://huggingface.co/spaces/open-
|
72 |
name: Open LLM Leaderboard
|
73 |
- task:
|
74 |
type: text-generation
|
@@ -171,9 +176,6 @@ The following hyperparameters were used during training:
|
|
171 |
- lr_scheduler_warmup_ratio: 0.1
|
172 |
- num_train_epochs: 6.0 (stop at epoch=1.0)
|
173 |
|
174 |
-
|
175 |
-
|
176 |
-
|
177 |
## Citation
|
178 |
```
|
179 |
@article{zhang2024general,
|
@@ -182,7 +184,4 @@ The following hyperparameters were used during training:
|
|
182 |
journal={arXiv preprint arXiv:2410.02197},
|
183 |
year={2024}
|
184 |
}
|
185 |
-
```
|
186 |
-
|
187 |
-
|
188 |
-
|
|
|
1 |
+
---
|
2 |
+
pipeline_tag: text-generation
|
3 |
+
library_name: transformers
|
4 |
+
license: apache-2.0
|
5 |
+
---
|
6 |
+
|
7 |
Quantization made by Richard Erkhov.
|
8 |
|
9 |
[Github](https://github.com/RichardErkhov)
|
|
|
12 |
|
13 |
[Request more models](https://github.com/RichardErkhov/quant_request)
|
14 |
|
15 |
+
Paper: [Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment](https://huggingface.co/papers/2410.02197)
|
16 |
+
Code: https://github.com/general-preference/general-preference-model
|
17 |
|
18 |
SPPO-Llama-3-8B-Instruct-GPM-2B - bnb 8bits
|
19 |
- Model creator: https://huggingface.co/general-preference/
|
20 |
- Original model: https://huggingface.co/general-preference/SPPO-Llama-3-8B-Instruct-GPM-2B/
|
21 |
|
|
|
|
|
|
|
22 |
Original model description:
|
23 |
---
|
24 |
language:
|
|
|
73 |
value: 8.01
|
74 |
name: exact match
|
75 |
source:
|
76 |
+
url: https://huggingface.co/spaces/open-llm_leaderboard/open_llm_leaderboard?query=general-preference/SPPO-Llama-3-8B-Instruct-GPM-2B
|
77 |
name: Open LLM Leaderboard
|
78 |
- task:
|
79 |
type: text-generation
|
|
|
176 |
- lr_scheduler_warmup_ratio: 0.1
|
177 |
- num_train_epochs: 6.0 (stop at epoch=1.0)
|
178 |
|
|
|
|
|
|
|
179 |
## Citation
|
180 |
```
|
181 |
@article{zhang2024general,
|
|
|
184 |
journal={arXiv preprint arXiv:2410.02197},
|
185 |
year={2024}
|
186 |
}
|
187 |
+
```
|
|
|
|
|
|