Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -69,7 +69,7 @@ tokenizer.pad_token = tokenizer.eos_token
|
|
69 |
model = AutoModelForCausalLM.from_pretrained(
|
70 |
MODEL_ID,
|
71 |
device_map="auto",
|
72 |
-
|
73 |
attn_implementation="flash_attention_2",
|
74 |
)
|
75 |
|
@@ -144,7 +144,8 @@ def predict(message, history, system_prompt, temperature, max_new_tokens, top_k,
|
|
144 |
'repetition_penalty':repetition_penalty,
|
145 |
'top_p':top_p,
|
146 |
'instruction':instruction,
|
147 |
-
'output':"".join(outputs)
|
|
|
148 |
}
|
149 |
hfapi.upload_file(
|
150 |
path_or_fileobj=json.dumps(dd, indent=2, ensure_ascii=False).encode('utf-8'),
|
|
|
69 |
model = AutoModelForCausalLM.from_pretrained(
|
70 |
MODEL_ID,
|
71 |
device_map="auto",
|
72 |
+
torch_dtype='auto',
|
73 |
attn_implementation="flash_attention_2",
|
74 |
)
|
75 |
|
|
|
144 |
'repetition_penalty':repetition_penalty,
|
145 |
'top_p':top_p,
|
146 |
'instruction':instruction,
|
147 |
+
'output':"".join(outputs),
|
148 |
+
'precision': 'auto '+str(model.dtype),
|
149 |
}
|
150 |
hfapi.upload_file(
|
151 |
path_or_fileobj=json.dumps(dd, indent=2, ensure_ascii=False).encode('utf-8'),
|