nielsr HF Staff commited on
Commit
b7b5c01
·
verified ·
1 Parent(s): ffa583f

Add link to GitHub repository

Browse files

This PR adds a direct link to the GitHub repository in the model card content for easier access to the code and further resources.

Files changed (1) hide show
  1. README.md +13 -10
README.md CHANGED
@@ -1,5 +1,11 @@
1
  ---
 
 
 
 
2
  license: cc-by-nc-4.0
 
 
3
  tags:
4
  - small-language-model
5
  - jee
@@ -12,14 +18,8 @@ tags:
12
  - mathematics
13
  - ai4education
14
  - physicswallah
15
- language:
16
- - en
17
- model_name: PhysicsWallah/Aryabhata-1.0
18
  model_creator: Physics Wallah AI Research
19
  model_type: Causal decoder-based model
20
- base_model: Qwen/Qwen2.5-Math-7B
21
- pipeline_tag: text-generation
22
- library_name: transformers
23
  ---
24
 
25
  # Aryabhatta 1.0 : An exam-focused language model for JEE Math
@@ -146,7 +146,8 @@ model = AutoModelForCausalLM.from_pretrained(model_id)
146
 
147
 
148
  # Define stop strings
149
- stop_strings = ["<|im_end|>", "<|end|>", "<im_start|>", "⁠```python\n", "⁠<|im_start|>", "]}}]}}]"]
 
150
 
151
  def strip_bad_tokens(s, stop_strings):
152
  for suffix in stop_strings:
@@ -192,7 +193,8 @@ llm = LLM(model="PhysicsWallahAI/Aryabhata-1.0")
192
  query = 'Find all the values of \\sqrt[3]{1}'
193
  messages = [{'role': 'system', 'content': 'Think step-by-step; put only the final answer inside \\boxed{}.'},
194
  {'role': 'user', 'content': query}]
195
- sampling_params = SamplingParams(temperature=0.0, max_tokens=4*1024, stop=["<|im_end|>", "<|end|>", "<im_start|>", "⁠```python\n", "⁠<|im_start|>", "]}}]}}]"])
 
196
 
197
  # Run inference
198
  results = llm.chat(messages, sampling_params)
@@ -203,7 +205,7 @@ print(results[0].outputs[0].text.strip())
203
 
204
  ---
205
 
206
- Read more about Aryabhata 1.0 in our [Technical Report](https://arxiv.org/abs/2508.08665)
207
 
208
  ---
209
 
@@ -226,4 +228,5 @@ If you use this model, please cite:
226
  author = {Physics Wallah AI Research},
227
  year = {2025},
228
  note = {\url{https://huggingface.co/PhysicsWallahAI/Aryabhata-1.0}},
229
- }
 
 
1
  ---
2
+ base_model: Qwen/Qwen2.5-Math-7B
3
+ language:
4
+ - en
5
+ library_name: transformers
6
  license: cc-by-nc-4.0
7
+ model_name: PhysicsWallah/Aryabhata-1.0
8
+ pipeline_tag: text-generation
9
  tags:
10
  - small-language-model
11
  - jee
 
18
  - mathematics
19
  - ai4education
20
  - physicswallah
 
 
 
21
  model_creator: Physics Wallah AI Research
22
  model_type: Causal decoder-based model
 
 
 
23
  ---
24
 
25
  # Aryabhatta 1.0 : An exam-focused language model for JEE Math
 
146
 
147
 
148
  # Define stop strings
149
+ stop_strings = ["<|im_end|>", "<|end|>", "<im_start|>", "⁠```python
150
+ ", "⁠<|im_start|>", "]}}]}}]"]
151
 
152
  def strip_bad_tokens(s, stop_strings):
153
  for suffix in stop_strings:
 
193
  query = 'Find all the values of \\sqrt[3]{1}'
194
  messages = [{'role': 'system', 'content': 'Think step-by-step; put only the final answer inside \\boxed{}.'},
195
  {'role': 'user', 'content': query}]
196
+ sampling_params = SamplingParams(temperature=0.0, max_tokens=4*1024, stop=["<|im_end|>", "<|end|>", "<im_start|>", "⁠```python
197
+ ", "⁠<|im_start|>", "]}}]}}]"])
198
 
199
  # Run inference
200
  results = llm.chat(messages, sampling_params)
 
205
 
206
  ---
207
 
208
+ Read more about Aryabhata 1.0 in our [Technical Report](https://arxiv.org/abs/2508.08665) and find the code on [GitHub](https://github.com/PhysicsWallahAI/Aryabhata-1.0).
209
 
210
  ---
211
 
 
228
  author = {Physics Wallah AI Research},
229
  year = {2025},
230
  note = {\url{https://huggingface.co/PhysicsWallahAI/Aryabhata-1.0}},
231
+ }
232
+ ```