Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,10 +1,9 @@
|
|
| 1 |
import gradio as gr
|
| 2 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
| 3 |
|
| 4 |
-
MODEL_NAME = "
|
| 5 |
-
|
| 6 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
| 7 |
-
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, torch_dtype="
|
| 8 |
generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
| 9 |
|
| 10 |
def chat(prompt, max_new_tokens=256, temperature=0.7):
|
|
@@ -16,7 +15,7 @@ def chat(prompt, max_new_tokens=256, temperature=0.7):
|
|
| 16 |
pad_token_id=tokenizer.eos_token_id
|
| 17 |
)
|
| 18 |
return outputs[0]['generated_text'][len(prompt):].strip()
|
| 19 |
-
|
| 20 |
iface = gr.Interface(
|
| 21 |
fn=chat,
|
| 22 |
inputs=[
|
|
|
|
| 1 |
import gradio as gr
|
| 2 |
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
| 3 |
|
| 4 |
+
MODEL_NAME = "openpeerai/openpeerllm"
|
|
|
|
| 5 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
| 6 |
+
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, torch_dtype="auto")
|
| 7 |
generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
| 8 |
|
| 9 |
def chat(prompt, max_new_tokens=256, temperature=0.7):
|
|
|
|
| 15 |
pad_token_id=tokenizer.eos_token_id
|
| 16 |
)
|
| 17 |
return outputs[0]['generated_text'][len(prompt):].strip()
|
| 18 |
+
MODEL_NAME = "openpeerai/openpeerllm"
|
| 19 |
iface = gr.Interface(
|
| 20 |
fn=chat,
|
| 21 |
inputs=[
|