Dan Flower commited on
Commit
aaa292f
·
1 Parent(s): d6b72de

model download issues

Browse files
Files changed (1) hide show
  1. model/model_runner.py +10 -1
model/model_runner.py CHANGED
@@ -35,7 +35,7 @@ def load_model():
35
 
36
  def generate(prompt: str, max_tokens: int = 256) -> str:
37
  model = load_model()
38
- response = model(
39
  prompt,
40
  max_tokens=max_tokens,
41
  stop=["\nUser:", "\nAssistant:"],
@@ -43,4 +43,13 @@ def generate(prompt: str, max_tokens: int = 256) -> str:
43
  temperature=0.7,
44
  top_p=0.95,
45
  )
 
 
 
 
 
 
 
 
 
46
  return response["choices"][0]["text"].strip()
 
35
 
36
  def generate(prompt: str, max_tokens: int = 256) -> str:
37
  model = load_model()
38
+ output = model(
39
  prompt,
40
  max_tokens=max_tokens,
41
  stop=["\nUser:", "\nAssistant:"],
 
43
  temperature=0.7,
44
  top_p=0.95,
45
  )
46
+ raw_text = output["choices"][0]["text"]
47
+
48
+ # Truncate at first stop sequence manually
49
+ for stop_token in ["\nUser:", "\nAssistant:"]:
50
+ if stop_token in raw_text:
51
+ raw_text = raw_text.split(stop_token)[0]
52
+
53
+ return raw_text.strip()
54
+
55
  return response["choices"][0]["text"].strip()