Spaces:
Sleeping
Sleeping
Dan Flower
commited on
Commit
·
aaa292f
1
Parent(s):
d6b72de
model download issues
Browse files- model/model_runner.py +10 -1
model/model_runner.py
CHANGED
@@ -35,7 +35,7 @@ def load_model():
|
|
35 |
|
36 |
def generate(prompt: str, max_tokens: int = 256) -> str:
|
37 |
model = load_model()
|
38 |
-
|
39 |
prompt,
|
40 |
max_tokens=max_tokens,
|
41 |
stop=["\nUser:", "\nAssistant:"],
|
@@ -43,4 +43,13 @@ def generate(prompt: str, max_tokens: int = 256) -> str:
|
|
43 |
temperature=0.7,
|
44 |
top_p=0.95,
|
45 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
46 |
return response["choices"][0]["text"].strip()
|
|
|
35 |
|
36 |
def generate(prompt: str, max_tokens: int = 256) -> str:
|
37 |
model = load_model()
|
38 |
+
output = model(
|
39 |
prompt,
|
40 |
max_tokens=max_tokens,
|
41 |
stop=["\nUser:", "\nAssistant:"],
|
|
|
43 |
temperature=0.7,
|
44 |
top_p=0.95,
|
45 |
)
|
46 |
+
raw_text = output["choices"][0]["text"]
|
47 |
+
|
48 |
+
# Truncate at first stop sequence manually
|
49 |
+
for stop_token in ["\nUser:", "\nAssistant:"]:
|
50 |
+
if stop_token in raw_text:
|
51 |
+
raw_text = raw_text.split(stop_token)[0]
|
52 |
+
|
53 |
+
return raw_text.strip()
|
54 |
+
|
55 |
return response["choices"][0]["text"].strip()
|