Update app.py
Browse files
app.py
CHANGED
|
@@ -44,14 +44,14 @@ def load_fn(models):
|
|
| 44 |
if model not in models_load.keys():
|
| 45 |
try:
|
| 46 |
m = gr.load(f'models/{model}')
|
| 47 |
-
print(f"{m}\n");
|
| 48 |
models_load.update({model: m})
|
| 49 |
models_load[model] = m # Store in dictionary
|
| 50 |
except Exception as error:
|
| 51 |
-
print(f"Error loading model {model}: {error}\n")
|
| 52 |
#m = gr.Interface(lambda _: None, inputs=gr.Textbox(), outputs=gr.Image(), queue=False)
|
| 53 |
#models_load.update({model: m})
|
| 54 |
-
traceback.print_exc() # Prints full error stack trace for debugging
|
| 55 |
#m = gr.Interface(fn=lambda _: None, inputs=gr.Textbox(), outputs=gr.Image(), queue=False)
|
| 56 |
models_load[model] = None
|
| 57 |
|
|
@@ -97,41 +97,41 @@ def gen_fn(model_str, prompt):
|
|
| 97 |
return None
|
| 98 |
elif isinstance(response, str):
|
| 99 |
if processed_models_count == 0:
|
| 100 |
-
print(f"***a***********")
|
| 101 |
# print(f"{prompt}")
|
| 102 |
-
print(f"{prompt}")
|
| 103 |
# print(f"{prompt}")
|
| 104 |
-
print(f"***b***********")
|
| 105 |
model_scores[current_model_name] += 1
|
| 106 |
-
print(f"OOO n:{processed_models_count} x:{current_model_index} r[{model_scores[current_model_name]}] {model_str}")
|
| 107 |
processed_models_count += 1
|
| 108 |
if processed_models_count == len(models):
|
| 109 |
-
|
| 110 |
-
|
| 111 |
processed_models_count = 0
|
| 112 |
return response
|
| 113 |
|
| 114 |
except TimeoutError:
|
| 115 |
-
print(f"TimeoutError: Model '{model_str}' did not respond within 150 seconds.")
|
| 116 |
processed_models_count += 1
|
| 117 |
if processed_models_count == len(models):
|
| 118 |
-
|
| 119 |
-
|
| 120 |
processed_models_count = 0
|
| 121 |
return None
|
| 122 |
|
| 123 |
except Exception as e:
|
| 124 |
if processed_models_count == 0:
|
| 125 |
-
|
| 126 |
# print(f"{prompt}")
|
| 127 |
# print(f"{prompt}")
|
| 128 |
# print(f"{prompt}")
|
| 129 |
-
|
| 130 |
-
print(f"--- n:{processed_models_count} x:{current_model_index} r[{model_scores[current_model_name]}] {model_str}")
|
| 131 |
processed_models_count += 1
|
| 132 |
if processed_models_count == len(models):
|
| 133 |
-
|
| 134 |
-
|
| 135 |
processed_models_count = 0
|
| 136 |
return None
|
| 137 |
|
|
|
|
| 44 |
if model not in models_load.keys():
|
| 45 |
try:
|
| 46 |
m = gr.load(f'models/{model}')
|
| 47 |
+
##print(f"{m}\n");
|
| 48 |
models_load.update({model: m})
|
| 49 |
models_load[model] = m # Store in dictionary
|
| 50 |
except Exception as error:
|
| 51 |
+
##print(f"Error loading model {model}: {error}\n")
|
| 52 |
#m = gr.Interface(lambda _: None, inputs=gr.Textbox(), outputs=gr.Image(), queue=False)
|
| 53 |
#models_load.update({model: m})
|
| 54 |
+
##traceback.print_exc() # Prints full error stack trace for debugging
|
| 55 |
#m = gr.Interface(fn=lambda _: None, inputs=gr.Textbox(), outputs=gr.Image(), queue=False)
|
| 56 |
models_load[model] = None
|
| 57 |
|
|
|
|
| 97 |
return None
|
| 98 |
elif isinstance(response, str):
|
| 99 |
if processed_models_count == 0:
|
| 100 |
+
##print(f"***a***********")
|
| 101 |
# print(f"{prompt}")
|
| 102 |
+
##print(f"{prompt}")
|
| 103 |
# print(f"{prompt}")
|
| 104 |
+
##print(f"***b***********")
|
| 105 |
model_scores[current_model_name] += 1
|
| 106 |
+
##print(f"OOO n:{processed_models_count} x:{current_model_index} r[{model_scores[current_model_name]}] {model_str}")
|
| 107 |
processed_models_count += 1
|
| 108 |
if processed_models_count == len(models):
|
| 109 |
+
## print("\nCycle Complete! Updated Scores:")
|
| 110 |
+
## print(model_scores)
|
| 111 |
processed_models_count = 0
|
| 112 |
return response
|
| 113 |
|
| 114 |
except TimeoutError:
|
| 115 |
+
##print(f"TimeoutError: Model '{model_str}' did not respond within 150 seconds.")
|
| 116 |
processed_models_count += 1
|
| 117 |
if processed_models_count == len(models):
|
| 118 |
+
## print("\nCycle Complete! Updated Scores:")
|
| 119 |
+
## print(model_scores)
|
| 120 |
processed_models_count = 0
|
| 121 |
return None
|
| 122 |
|
| 123 |
except Exception as e:
|
| 124 |
if processed_models_count == 0:
|
| 125 |
+
## print(f"******c*******")
|
| 126 |
# print(f"{prompt}")
|
| 127 |
# print(f"{prompt}")
|
| 128 |
# print(f"{prompt}")
|
| 129 |
+
## print(f"******d*******")
|
| 130 |
+
##print(f"--- n:{processed_models_count} x:{current_model_index} r[{model_scores[current_model_name]}] {model_str}")
|
| 131 |
processed_models_count += 1
|
| 132 |
if processed_models_count == len(models):
|
| 133 |
+
## print("\nCycle Complete! Updated Scores:")
|
| 134 |
+
## print(model_scores)
|
| 135 |
processed_models_count = 0
|
| 136 |
return None
|
| 137 |
|