Improve the Metharme prompt

This commit is contained in:
oobabooga 2023-05-10 16:09:32 -03:00
parent e5b1547849
commit b7a589afc8
2 changed files with 7 additions and 5 deletions

View File

@ -1,4 +1,4 @@
user: "<|user|>"
bot: "<|model|>"
context: "<|system|>"
context: "<|system|>This is a text adventure game. Describe the scenario to the user and give him three options to pick from on each turn."
turn_template: "<|user|><|user-message|><|bot|><|bot-message|>"

View File

@ -188,7 +188,7 @@ def chatbot_wrapper(text, state, regenerate=False, _continue=False):
# Generate
for i in range(state['chat_generation_attempts']):
reply = None
for reply in generate_reply(f"{prompt}{' ' if len(cumulative_reply) > 0 else ''}{cumulative_reply}", state, eos_token=eos_token, stopping_strings=stopping_strings):
for j, reply in enumerate(generate_reply(f"{prompt}{' ' if len(cumulative_reply) > 0 else ''}{cumulative_reply}", state, eos_token=eos_token, stopping_strings=stopping_strings)):
reply = cumulative_reply + reply
# Extracting the reply
@ -212,9 +212,11 @@ def chatbot_wrapper(text, state, regenerate=False, _continue=False):
shared.history['internal'].append(['', ''])
shared.history['visible'].append(['', ''])
if not (j == 0 and visible_reply.strip() == ''):
shared.history['internal'][-1] = [text, reply]
shared.history['visible'][-1] = [visible_text, visible_reply]
yield shared.history['visible']
if next_character_found:
break