From 3f1e70d2c8e385d15f52dfceef374389e4ba548e Mon Sep 17 00:00:00 2001 From: oobabooga <112222186+oobabooga@users.noreply.github.com> Date: Fri, 13 Jan 2023 14:00:43 -0300 Subject: [PATCH] Remove the temperature slider It was not being used by most presets. --- presets/Naive.txt | 2 +- server.py | 35 ++++++++++++++++------------------- 2 files changed, 17 insertions(+), 20 deletions(-) diff --git a/presets/Naive.txt b/presets/Naive.txt index 1f66840c..e323a580 100644 --- a/presets/Naive.txt +++ b/presets/Naive.txt @@ -2,4 +2,4 @@ do_sample=True, max_new_tokens=max_length, top_p=1, typical_p=0.3, -temperature=temperature, +temperature=0.7, diff --git a/server.py b/server.py index 77fcd7f8..af431801 100644 --- a/server.py +++ b/server.py @@ -94,7 +94,7 @@ def generate_html(s): s = f'
{s}
' return s -def generate_reply(question, temperature, max_length, inference_settings, selected_model, eos_token=None): +def generate_reply(question, max_length, inference_settings, selected_model, eos_token=None): global model, tokenizer, model_name, loaded_preset, preset if selected_model != model_name: @@ -179,17 +179,16 @@ if args.notebook: with gr.Row(): with gr.Column(): length_slider = gr.Slider(minimum=1, maximum=2000, step=1, label='max_length', value=200) - temp_slider = gr.Slider(minimum=0.0, maximum=1.0, step=0.01, label='Temperature', value=0.7) with gr.Column(): - preset_menu = gr.Dropdown(choices=available_presets, value="NovelAI-Sphinx Moth", label='Preset') model_menu = gr.Dropdown(choices=available_models, value=model_name, label='Model') + preset_menu = gr.Dropdown(choices=available_presets, value="NovelAI-Sphinx Moth", label='Preset') - btn.click(generate_reply, [textbox, temp_slider, length_slider, preset_menu, model_menu], [textbox, markdown, html], show_progress=True, api_name="textgen") - textbox.submit(generate_reply, [textbox, temp_slider, length_slider, preset_menu, model_menu], [textbox, markdown, html], show_progress=True) + btn.click(generate_reply, [textbox, length_slider, preset_menu, model_menu], [textbox, markdown, html], show_progress=True, api_name="textgen") + textbox.submit(generate_reply, [textbox, length_slider, preset_menu, model_menu], [textbox, markdown, html], show_progress=True) elif args.chat: history = [] - def chatbot_wrapper(text, temperature, max_length, inference_settings, selected_model, name1, name2, context): + def chatbot_wrapper(text, max_length, inference_settings, selected_model, name1, name2, context): question = context+'\n\n' for i in range(len(history)): question += f"{name1}: {history[i][0][3:-5].strip()}\n" @@ -197,7 +196,7 @@ elif args.chat: question += f"{name1}: {text.strip()}\n" question += f"{name2}:" - reply = generate_reply(question, temperature, max_length, inference_settings, selected_model, eos_token='\n')[0] + reply = generate_reply(question, max_length, inference_settings, selected_model, eos_token='\n')[0] reply = reply[len(question):].split('\n')[0].strip() history.append((text, reply)) return history @@ -207,7 +206,7 @@ elif args.chat: history = [] if 'pygmalion' in model_name.lower(): - context_str = "Kawaii's Persona: Kawaii is an enthusiastic and nice girl.\n" + context_str = "This is a conversation between two people.\n" name1_str = "You" name2_str = "Kawaii" else: @@ -221,11 +220,10 @@ elif args.chat: with gr.Column(): with gr.Row(): with gr.Column(): - length_slider = gr.Slider(minimum=1, maximum=2000, step=1, label='max_length', value=200) + model_menu = gr.Dropdown(choices=available_models, value=model_name, label='Model') preset_menu = gr.Dropdown(choices=available_presets, value="NovelAI-Sphinx Moth", label='Preset') with gr.Column(): - temp_slider = gr.Slider(minimum=0.0, maximum=1.0, step=0.01, label='Temperature', value=0.7) - model_menu = gr.Dropdown(choices=available_models, value=model_name, label='Model') + length_slider = gr.Slider(minimum=1, maximum=2000, step=1, label='max_length', value=200) name1 = gr.Textbox(value=name1_str, lines=1, label='Your name') name2 = gr.Textbox(value=name2_str, lines=1, label='Bot\'s name') context = gr.Textbox(value=context_str, lines=2, label='Context') @@ -235,16 +233,16 @@ elif args.chat: btn = gr.Button("Generate") btn2 = gr.Button("Clear history") - btn.click(chatbot_wrapper, [textbox, temp_slider, length_slider, preset_menu, model_menu, name1, name2, context], display1, show_progress=True, api_name="textgen") - textbox.submit(chatbot_wrapper, [textbox, temp_slider, length_slider, preset_menu, model_menu, name1, name2, context], display1, show_progress=True) + btn.click(chatbot_wrapper, [textbox, length_slider, preset_menu, model_menu, name1, name2, context], display1, show_progress=True, api_name="textgen") + textbox.submit(chatbot_wrapper, [textbox, length_slider, preset_menu, model_menu, name1, name2, context], display1, show_progress=True) btn2.click(clear) btn.click(lambda x: "", textbox, textbox, show_progress=False) textbox.submit(lambda x: "", textbox, textbox, show_progress=False) btn2.click(lambda x: "", display1, display1) else: - def continue_wrapper(question, temperature, max_length, inference_settings, selected_model): - a, b, c = generate_reply(question, temperature, max_length, inference_settings, selected_model) + def continue_wrapper(question, max_length, inference_settings, selected_model): + a, b, c = generate_reply(question, max_length, inference_settings, selected_model) return a, a, b, c with gr.Blocks(css=css, analytics_enabled=False) as interface: @@ -252,7 +250,6 @@ else: with gr.Row(): with gr.Column(): textbox = gr.Textbox(value=default_text, lines=15, label='Input') - temp_slider = gr.Slider(minimum=0.0, maximum=1.0, step=0.01, label='Temperature', value=0.7) length_slider = gr.Slider(minimum=1, maximum=2000, step=1, label='max_length', value=200) preset_menu = gr.Dropdown(choices=available_presets, value="NovelAI-Sphinx Moth", label='Preset') model_menu = gr.Dropdown(choices=available_models, value=model_name, label='Model') @@ -266,9 +263,9 @@ else: with gr.Tab('HTML'): html = gr.HTML() - btn.click(generate_reply, [textbox, temp_slider, length_slider, preset_menu, model_menu], [output_textbox, markdown, html], show_progress=True, api_name="textgen") - cont.click(continue_wrapper, [output_textbox, temp_slider, length_slider, preset_menu, model_menu], [output_textbox, textbox, markdown, html], show_progress=True) - textbox.submit(generate_reply, [textbox, temp_slider, length_slider, preset_menu, model_menu], [output_textbox, markdown, html], show_progress=True) + btn.click(generate_reply, [textbox, length_slider, preset_menu, model_menu], [output_textbox, markdown, html], show_progress=True, api_name="textgen") + cont.click(continue_wrapper, [output_textbox, length_slider, preset_menu, model_menu], [output_textbox, textbox, markdown, html], show_progress=True) + textbox.submit(generate_reply, [textbox, length_slider, preset_menu, model_menu], [output_textbox, markdown, html], show_progress=True) if args.no_listen: interface.launch(share=False)