model-evaluation / tabs /arena_side_by_side.py
AlekseyKorshuk's picture
updates
e053720
import time
import gradio as gr
import random
from conversation import Conversation
from utils import get_matchmaking
def get_tab_arena_side_by_side(download_bot_config, get_bot_profile, model_mapping, client):
gr.Markdown("""
# ⚔️ Chatbot Arena (side-by-side) ⚔️
## Rules
* Chat with two models side-by-side and vote for which one is better!
* You pick the models you want to chat with.
* You can continue chatting and voting or click “Clear” to start a new round.
""")
default_bot_id = "_bot_e21de304-6151-4a04-b025-4c553ae8cbca"
bot_config = download_bot_config(default_bot_id)
user_state = gr.State(
bot_config
)
with gr.Row():
bot_id = gr.Textbox(label="Chai bot ID", value=default_bot_id, interactive=True)
reload_bot_button = gr.Button("Reload bot")
bot_profile = gr.HTML(get_bot_profile(bot_config))
with gr.Accordion("Bot config:", open=False):
bot_config_text = gr.Markdown(f"# Memory\n{bot_config['memory']}\n# Prompt\n{bot_config['prompt']}\n")
with gr.Row():
values = list(model_mapping.keys())
first_message = (None, bot_config["firstMessage"])
height = 450
model_a_value, model_b_value = get_matchmaking(client, values, is_anonymous=False)
with gr.Column():
model_a = gr.Dropdown(values, value=model_a_value, label="Model A")
chatbot_a = gr.Chatbot([first_message])
chatbot_a.style(height=height)
with gr.Column():
model_b = gr.Dropdown(values, value=model_b_value, label="Model B")
chatbot_b = gr.Chatbot([first_message])
chatbot_b.style(height=height)
with gr.Row():
with gr.Column(scale=3):
msg = gr.Textbox(show_label=False, value="Hi there!", interactive=True)
with gr.Column(scale=3):
send = gr.Button("Send")
with gr.Row():
vote_a = gr.Button("👈 A is better", interactive=False)
vote_b = gr.Button("👉 B is better", interactive=False)
vote_tie = gr.Button("🤝 Tie", interactive=False)
vote_bad = gr.Button("💩 Both are bad", interactive=False)
with gr.Row():
regenerate = gr.Button("Regenerate", interactive=False)
clear = gr.Button("Clear")
with gr.Accordion("Generation parameters for model A", open=False):
model = model_mapping[model_a.value]
temperature_model_a = gr.Slider(minimum=0.0, maximum=1.0, value=model.generation_params["temperature"],
interactive=True, label="Temperature")
repetition_penalty_model_a = gr.Slider(minimum=0.0, maximum=2.0,
value=model.generation_params["repetition_penalty"],
interactive=True, label="Repetition penalty")
max_new_tokens_model_a = gr.Slider(minimum=1, maximum=512, value=model.generation_params["max_new_tokens"],
interactive=True, label="Max new tokens")
top_k_model_a = gr.Slider(minimum=1, maximum=100, value=model.generation_params["top_k"],
interactive=True, label="Top-K")
top_p_model_a = gr.Slider(minimum=0.0, maximum=1.0, value=model.generation_params["top_p"],
interactive=True, label="Top-P")
with gr.Accordion("Generation parameters for model B", open=False):
model = model_mapping[model_b.value]
temperature_model_b = gr.Slider(minimum=0.0, maximum=1.0, value=model.generation_params["temperature"],
interactive=True, label="Temperature")
repetition_penalty_model_b = gr.Slider(minimum=0.0, maximum=2.0,
value=model.generation_params["repetition_penalty"],
interactive=True, label="Repetition penalty")
max_new_tokens_model_b = gr.Slider(minimum=1, maximum=512, value=model.generation_params["max_new_tokens"],
interactive=True, label="Max new tokens")
top_k_model_b = gr.Slider(minimum=1, maximum=100, value=model.generation_params["top_k"],
interactive=True, label="Top-K")
top_p_model_b = gr.Slider(minimum=0.0, maximum=1.0, value=model.generation_params["top_p"],
interactive=True, label="Top-P")
def clear_chat(user_state):
return "", [(None, user_state["firstMessage"])], [(None, user_state["firstMessage"])]
def reload_bot(bot_id):
bot_config = download_bot_config(bot_id)
bot_profile = get_bot_profile(bot_config)
return bot_profile, [(None, bot_config["firstMessage"])], [(None, bot_config[
"firstMessage"])], bot_config, f"# Memory\n{bot_config['memory']}\n# Prompt\n{bot_config['prompt']}"
def get_generation_args(model_tag):
model = model_mapping[model_tag]
return (
model.generation_params["temperature"],
model.generation_params["repetition_penalty"],
model.generation_params["max_new_tokens"],
model.generation_params["top_k"],
model.generation_params["top_p"],
)
def respond(message, chat_history, user_state, model_tag,
temperature, repetition_penalty, max_new_tokens, top_k, top_p):
custom_generation_params = {
'temperature': temperature,
'repetition_penalty': repetition_penalty,
'max_new_tokens': max_new_tokens,
'top_k': top_k,
'top_p': top_p,
}
conv = Conversation(user_state)
conv.set_chat_history(chat_history)
conv.add_user_message(message)
model = model_mapping[model_tag]
bot_message = model.generate_response(conv, custom_generation_params)
chat_history.append(
(message, bot_message)
)
return "", chat_history
def record_vote(user_state, vote,
chat_history_a, model_tag_a,
chat_history_b, model_tag_b):
if len(chat_history_a) < 2:
return
conv_a = Conversation(user_state)
conv_a.set_chat_history(chat_history_a)
conv_b = Conversation(user_state)
conv_b.set_chat_history(chat_history_b)
if "A is better" in vote:
vote_str = "model_a"
elif "B is better" in vote:
vote_str = "model_b"
elif "Tie" in vote:
vote_str = "tie"
else:
vote_str = "tie (bothbad)"
row = {
"timestamp": time.time(),
"bot_id": user_state["bot_id"],
"vote": vote_str,
"model_a": model_tag_a,
"model_b": model_tag_b,
"is_anonymous": int(False)
}
sheet = client.open("Chat Arena").sheet1
num_rows = len(sheet.get_all_records())
sheet.insert_row(list(row.values()), index=num_rows + 2)
return
def regenerate_response(chat_history, user_state, model_tag,
temperature, repetition_penalty, max_new_tokens, top_k, top_p):
custom_generation_params = {
'temperature': temperature,
'repetition_penalty': repetition_penalty,
'max_new_tokens': max_new_tokens,
'top_k': top_k,
'top_p': top_p,
}
last_row = chat_history.pop(-1)
chat_history.append((last_row[0], None))
model = model_mapping[model_tag]
conv = Conversation(user_state)
conv.set_chat_history(chat_history)
bot_message = model.generate_response(conv, custom_generation_params)
chat_history[-1] = (last_row[0], bot_message)
return "", chat_history
def disable_voting():
return [gr.Button.update(interactive=False)] * 4
def enable_voting():
return [gr.Button.update(interactive=True)] * 4
def enable_send():
return [gr.Button.update(interactive=True), gr.Button.update(interactive=False)]
def enable_regenerate():
return gr.Button.update(interactive=True)
for vote in [vote_a, vote_b, vote_tie, vote_bad]:
vote.click(record_vote,
[user_state, vote, chatbot_a, model_a, chatbot_b, model_b],
None,
queue=False)
vote.click(disable_voting, None, [vote_a, vote_b, vote_tie, vote_bad], queue=False)
model_a.change(get_generation_args, [model_a],
[temperature_model_a, repetition_penalty_model_a, max_new_tokens_model_a, top_k_model_a,
top_p_model_a], queue=False)
model_b.change(get_generation_args, [model_b],
[temperature_model_b, repetition_penalty_model_b, max_new_tokens_model_b, top_k_model_b,
top_p_model_b], queue=False)
reload_bot_button.click(reload_bot, [bot_id], [bot_profile, chatbot_a, chatbot_b, user_state, bot_config_text],
queue=False)
clear.click(clear_chat, [user_state], [msg, chatbot_a, chatbot_b], queue=False)
model_a.change(clear_chat, [user_state], [msg, chatbot_a, chatbot_b], queue=False)
model_b.change(clear_chat, [user_state], [msg, chatbot_a, chatbot_b], queue=False)
clear.click(enable_send, None, [send, regenerate], queue=False)
reload_bot_button.click(enable_send, None, [send, regenerate], queue=False)
model_a.change(enable_voting, None, [vote_a, vote_b, vote_tie, vote_bad], queue=False)
model_b.change(enable_voting, None, [vote_a, vote_b, vote_tie, vote_bad], queue=False)
reload_bot_button.click(disable_voting, None, [vote_a, vote_b, vote_tie, vote_bad], queue=False)
send.click(enable_voting, None, [vote_a, vote_b, vote_tie, vote_bad], queue=False)
clear.click(disable_voting, None, [vote_a, vote_b, vote_tie, vote_bad], queue=False)
regenerate.click(enable_voting, None, [vote_a, vote_b, vote_tie, vote_bad], queue=False)
msg.submit(enable_voting, None, [vote_a, vote_b, vote_tie, vote_bad], queue=False)
send.click(respond,
[msg, chatbot_a, user_state, model_a, temperature_model_a, repetition_penalty_model_a,
max_new_tokens_model_a, top_k_model_a, top_p_model_a], [msg, chatbot_a],
queue=False)
msg.submit(respond,
[msg, chatbot_a, user_state, model_a, temperature_model_a, repetition_penalty_model_a,
max_new_tokens_model_a, top_k_model_a, top_p_model_a], [msg, chatbot_a],
queue=False)
send.click(respond,
[msg, chatbot_b, user_state, model_b, temperature_model_b, repetition_penalty_model_b,
max_new_tokens_model_b, top_k_model_b, top_p_model_b], [msg, chatbot_b],
queue=False)
msg.submit(respond,
[msg, chatbot_b, user_state, model_b, temperature_model_b, repetition_penalty_model_b,
max_new_tokens_model_b, top_k_model_b, top_p_model_b], [msg, chatbot_b],
queue=False)
send.click(enable_regenerate, None, [regenerate], queue=False)
msg.submit(enable_regenerate, None, [regenerate], queue=False)
regenerate.click(regenerate_response,
[chatbot_a, user_state, model_a, temperature_model_a, repetition_penalty_model_a,
max_new_tokens_model_a, top_k_model_a,
top_p_model_a], [msg, chatbot_a], queue=False)
regenerate.click(regenerate_response,
[chatbot_b, user_state, model_b, temperature_model_b, repetition_penalty_model_b,
max_new_tokens_model_b, top_k_model_b,
top_p_model_b], [msg, chatbot_b], queue=False)