mirror of
https://github.com/abetlen/llama-cpp-python.git
synced 2023-09-07 17:34:22 +03:00
Bugfix
This commit is contained in:
@@ -695,10 +695,7 @@ class Llama:
|
|||||||
Returns:
|
Returns:
|
||||||
Generated chat completion or a stream of chat completion chunks.
|
Generated chat completion or a stream of chat completion chunks.
|
||||||
"""
|
"""
|
||||||
stop = stop if not None else []
|
stop = stop if stop is not None else []
|
||||||
instructions = """Complete the following chat conversation between the user and the assistant. System messages should be strictly followed as additional instructions."""
|
|
||||||
chat_history = "\n".join(
|
|
||||||
f'{message["role"]} {message.get("user", "")}: {message["content"]}'
|
|
||||||
for message in messages
|
for message in messages
|
||||||
)
|
)
|
||||||
PROMPT = f" \n\n### Instructions:{instructions}\n\n### Inputs:{chat_history}\n\n### Response:\nassistant: "
|
PROMPT = f" \n\n### Instructions:{instructions}\n\n### Inputs:{chat_history}\n\n### Response:\nassistant: "
|
||||||
|
|||||||
Reference in New Issue
Block a user