Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -30,6 +30,7 @@ MODEL_PATH = str(mistral_models_path)
|
|
| 30 |
tokenizer = MistralTokenizer.from_file(os.path.join(MODEL_PATH, "tokenizer.model.v3"))
|
| 31 |
model = Transformer.from_folder(MODEL_PATH)
|
| 32 |
|
|
|
|
| 33 |
@spaces.GPU()
|
| 34 |
def generate_response(message, history):
|
| 35 |
# Convert history to the format expected by the model
|
|
@@ -46,13 +47,14 @@ def generate_response(message, history):
|
|
| 46 |
tokens = tokenizer.encode_chat_completion(completion_request).tokens
|
| 47 |
|
| 48 |
# Generate response
|
| 49 |
-
out_tokens
|
| 50 |
|
| 51 |
# Decode response
|
| 52 |
result = tokenizer.instruct_tokenizer.tokenizer.decode(out_tokens[0])
|
| 53 |
|
| 54 |
return result
|
| 55 |
|
|
|
|
| 56 |
# Gradio interface
|
| 57 |
iface = gr.ChatInterface(
|
| 58 |
generate_response,
|
|
|
|
| 30 |
tokenizer = MistralTokenizer.from_file(os.path.join(MODEL_PATH, "tokenizer.model.v3"))
|
| 31 |
model = Transformer.from_folder(MODEL_PATH)
|
| 32 |
|
| 33 |
+
|
| 34 |
@spaces.GPU()
|
| 35 |
def generate_response(message, history):
|
| 36 |
# Convert history to the format expected by the model
|
|
|
|
| 47 |
tokens = tokenizer.encode_chat_completion(completion_request).tokens
|
| 48 |
|
| 49 |
# Generate response
|
| 50 |
+
out_tokens = generate([tokens], model, max_tokens=256, temperature=0.7, eos_id=tokenizer.instruct_tokenizer.tokenizer.eos_id)
|
| 51 |
|
| 52 |
# Decode response
|
| 53 |
result = tokenizer.instruct_tokenizer.tokenizer.decode(out_tokens[0])
|
| 54 |
|
| 55 |
return result
|
| 56 |
|
| 57 |
+
|
| 58 |
# Gradio interface
|
| 59 |
iface = gr.ChatInterface(
|
| 60 |
generate_response,
|