Spaces:
Runtime error
Runtime error
more ref, ci, debug
Browse files
app.py
CHANGED
|
@@ -28,6 +28,15 @@ from src.envs import API, EVAL_REQUESTS_PATH, EVAL_RESULTS_PATH, QUEUE_REPO, REP
|
|
| 28 |
from src.populate import get_evaluation_queue_df, get_leaderboard_df
|
| 29 |
from src.submission.submit import add_new_eval
|
| 30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
|
| 32 |
def restart_space():
|
| 33 |
API.restart_space(repo_id=REPO_ID)
|
|
@@ -94,6 +103,8 @@ demo = gr.Blocks(css=custom_css)
|
|
| 94 |
with demo:
|
| 95 |
gr.HTML(TITLE)
|
| 96 |
gr.Markdown(INTRODUCTION_TEXT, elem_classes="markdown-text")
|
|
|
|
|
|
|
| 97 |
|
| 98 |
with gr.Tabs(elem_classes="tab-buttons") as tabs:
|
| 99 |
with gr.TabItem("🏅 LLM Benchmark", elem_id="llm-benchmark-tab-table", id=0):
|
|
@@ -141,7 +152,7 @@ with demo:
|
|
| 141 |
# headers=EVAL_COLS,
|
| 142 |
# datatype=EVAL_TYPES,
|
| 143 |
# row_count=5,
|
| 144 |
-
|
| 145 |
with gr.Row():
|
| 146 |
gr.Markdown("# ✉️✨ Submit your model here!", elem_classes="markdown-text")
|
| 147 |
|
|
@@ -174,22 +185,25 @@ with demo:
|
|
| 174 |
)
|
| 175 |
base_model_name_textbox = gr.Textbox(label="Организация")
|
| 176 |
ans_file = gr.File(label="Arena Hard Answer File", file_types=["json","jsonl"])
|
| 177 |
-
|
| 178 |
-
|
| 179 |
-
|
| 180 |
-
|
| 181 |
-
|
| 182 |
-
|
| 183 |
-
|
| 184 |
-
|
| 185 |
-
|
| 186 |
-
|
| 187 |
-
|
| 188 |
-
|
| 189 |
-
|
| 190 |
-
|
| 191 |
-
|
| 192 |
-
|
|
|
|
|
|
|
|
|
|
| 193 |
|
| 194 |
with gr.Row():
|
| 195 |
with gr.Accordion("📙 Citation", open=False):
|
|
|
|
| 28 |
from src.populate import get_evaluation_queue_df, get_leaderboard_df
|
| 29 |
from src.submission.submit import add_new_eval
|
| 30 |
|
| 31 |
+
loggedin = False
|
| 32 |
+
|
| 33 |
+
def check_login(profile: gr.OAuthProfile | None) -> bool:
|
| 34 |
+
global loggedin
|
| 35 |
+
if profile is None:
|
| 36 |
+
loggedin = False
|
| 37 |
+
return False
|
| 38 |
+
loggedin = True
|
| 39 |
+
return True
|
| 40 |
|
| 41 |
def restart_space():
|
| 42 |
API.restart_space(repo_id=REPO_ID)
|
|
|
|
| 103 |
with demo:
|
| 104 |
gr.HTML(TITLE)
|
| 105 |
gr.Markdown(INTRODUCTION_TEXT, elem_classes="markdown-text")
|
| 106 |
+
login_button = gr.LoginButton(elem_id="oauth-button")
|
| 107 |
+
|
| 108 |
|
| 109 |
with gr.Tabs(elem_classes="tab-buttons") as tabs:
|
| 110 |
with gr.TabItem("🏅 LLM Benchmark", elem_id="llm-benchmark-tab-table", id=0):
|
|
|
|
| 152 |
# headers=EVAL_COLS,
|
| 153 |
# datatype=EVAL_TYPES,
|
| 154 |
# row_count=5,
|
| 155 |
+
|
| 156 |
with gr.Row():
|
| 157 |
gr.Markdown("# ✉️✨ Submit your model here!", elem_classes="markdown-text")
|
| 158 |
|
|
|
|
| 185 |
)
|
| 186 |
base_model_name_textbox = gr.Textbox(label="Организация")
|
| 187 |
ans_file = gr.File(label="Arena Hard Answer File", file_types=["json","jsonl"])
|
| 188 |
+
login_button.click(check_login)
|
| 189 |
+
if not loggedin:
|
| 190 |
+
gr.Markdown("Please log in to submit an evaluation.")
|
| 191 |
+
else:
|
| 192 |
+
submit_button = gr.Button("Submit Eval")
|
| 193 |
+
submission_result = gr.Markdown()
|
| 194 |
+
submit_button.click(
|
| 195 |
+
add_new_eval,
|
| 196 |
+
[
|
| 197 |
+
model_name_textbox,
|
| 198 |
+
base_model_name_textbox,
|
| 199 |
+
revision_name_textbox,
|
| 200 |
+
precision,
|
| 201 |
+
weight_type,
|
| 202 |
+
model_type,
|
| 203 |
+
ans_file,
|
| 204 |
+
],
|
| 205 |
+
submission_result,
|
| 206 |
+
)
|
| 207 |
|
| 208 |
with gr.Row():
|
| 209 |
with gr.Accordion("📙 Citation", open=False):
|