RichardErkhov nielsr HF Staff commited on
Commit
b6c0191
·
verified ·
1 Parent(s): d4816c0

Add model card metadata and link to paper and code repository (#1)

Browse files

- Add model card metadata and link to paper and code repository (ab146d1f70834483a042d8485d695be3d053cccc)


Co-authored-by: Niels Rogge <nielsr@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +10 -11
README.md CHANGED
@@ -1,3 +1,9 @@
 
 
 
 
 
 
1
  Quantization made by Richard Erkhov.
2
 
3
  [Github](https://github.com/RichardErkhov)
@@ -6,14 +12,13 @@ Quantization made by Richard Erkhov.
6
 
7
  [Request more models](https://github.com/RichardErkhov/quant_request)
8
 
 
 
9
 
10
  SPPO-Llama-3-8B-Instruct-GPM-2B - bnb 8bits
11
  - Model creator: https://huggingface.co/general-preference/
12
  - Original model: https://huggingface.co/general-preference/SPPO-Llama-3-8B-Instruct-GPM-2B/
13
 
14
-
15
-
16
-
17
  Original model description:
18
  ---
19
  language:
@@ -68,7 +73,7 @@ model-index:
68
  value: 8.01
69
  name: exact match
70
  source:
71
- url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=general-preference/SPPO-Llama-3-8B-Instruct-GPM-2B
72
  name: Open LLM Leaderboard
73
  - task:
74
  type: text-generation
@@ -171,9 +176,6 @@ The following hyperparameters were used during training:
171
  - lr_scheduler_warmup_ratio: 0.1
172
  - num_train_epochs: 6.0 (stop at epoch=1.0)
173
 
174
-
175
-
176
-
177
  ## Citation
178
  ```
179
  @article{zhang2024general,
@@ -182,7 +184,4 @@ The following hyperparameters were used during training:
182
  journal={arXiv preprint arXiv:2410.02197},
183
  year={2024}
184
  }
185
- ```
186
-
187
-
188
-
 
1
+ ---
2
+ pipeline_tag: text-generation
3
+ library_name: transformers
4
+ license: apache-2.0
5
+ ---
6
+
7
  Quantization made by Richard Erkhov.
8
 
9
  [Github](https://github.com/RichardErkhov)
 
12
 
13
  [Request more models](https://github.com/RichardErkhov/quant_request)
14
 
15
+ Paper: [Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment](https://huggingface.co/papers/2410.02197)
16
+ Code: https://github.com/general-preference/general-preference-model
17
 
18
  SPPO-Llama-3-8B-Instruct-GPM-2B - bnb 8bits
19
  - Model creator: https://huggingface.co/general-preference/
20
  - Original model: https://huggingface.co/general-preference/SPPO-Llama-3-8B-Instruct-GPM-2B/
21
 
 
 
 
22
  Original model description:
23
  ---
24
  language:
 
73
  value: 8.01
74
  name: exact match
75
  source:
76
+ url: https://huggingface.co/spaces/open-llm_leaderboard/open_llm_leaderboard?query=general-preference/SPPO-Llama-3-8B-Instruct-GPM-2B
77
  name: Open LLM Leaderboard
78
  - task:
79
  type: text-generation
 
176
  - lr_scheduler_warmup_ratio: 0.1
177
  - num_train_epochs: 6.0 (stop at epoch=1.0)
178
 
 
 
 
179
  ## Citation
180
  ```
181
  @article{zhang2024general,
 
184
  journal={arXiv preprint arXiv:2410.02197},
185
  year={2024}
186
  }
187
+ ```