Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -10,10 +10,10 @@ login(api_key)
|
|
10 |
# Predefined list of models to compare (can be expanded)
|
11 |
model_options = {
|
12 |
"Llama-3.1-70B": "meta-llama/Llama-3.1-70B-Instruct",
|
13 |
-
"
|
14 |
-
"
|
15 |
-
"
|
16 |
-
"
|
17 |
}
|
18 |
|
19 |
# Initialize clients for models
|
@@ -77,7 +77,7 @@ def create_demo():
|
|
77 |
selected_models = gr.CheckboxGroup(
|
78 |
choices=list(model_options.keys()),
|
79 |
label="Select models to compare",
|
80 |
-
value=["Llama-3.1-70B", "
|
81 |
)
|
82 |
|
83 |
submit_button = gr.Button("Generate Responses")
|
@@ -126,7 +126,7 @@ def create_demo():
|
|
126 |
outputs=[response[2] for response in response_boxes],
|
127 |
)
|
128 |
|
129 |
-
vote_state = gr.State([0] * len(model_options)) #
|
130 |
|
131 |
for model_name, _, _, vote_button, vote_count in response_boxes:
|
132 |
vote_button.click(
|
|
|
10 |
# Predefined list of models to compare (can be expanded)
|
11 |
model_options = {
|
12 |
"Llama-3.1-70B": "meta-llama/Llama-3.1-70B-Instruct",
|
13 |
+
"Qwen-2.5-1.5B-Instruct": "Qwen/Qwen2.5-1.5B-Instruct",
|
14 |
+
"Llama-3.2-1B": "meta-llama/Llama-3.2-1B",
|
15 |
+
"DeepSeek-V2.5": "deepseek-ai/DeepSeek-V2.5",
|
16 |
+
"Athene-V2-Chat": "Nexusflow/Athene-V2-Chat",
|
17 |
}
|
18 |
|
19 |
# Initialize clients for models
|
|
|
77 |
selected_models = gr.CheckboxGroup(
|
78 |
choices=list(model_options.keys()),
|
79 |
label="Select models to compare",
|
80 |
+
value=["Llama-3.1-70B", "Qwen-2.5-1.5B-Instruct"], # Default models
|
81 |
)
|
82 |
|
83 |
submit_button = gr.Button("Generate Responses")
|
|
|
126 |
outputs=[response[2] for response in response_boxes],
|
127 |
)
|
128 |
|
129 |
+
vote_state = gr.State([0] * len(model_options)) # Initialize one vote count per model
|
130 |
|
131 |
for model_name, _, _, vote_button, vote_count in response_boxes:
|
132 |
vote_button.click(
|