Spaces:
Sleeping
Sleeping
add output for step 7
Browse files
app.py
CHANGED
|
@@ -235,12 +235,13 @@ def get_prefinal_response(row):
|
|
| 235 |
print('{}'.format(modified_query))
|
| 236 |
print('\nStep 7: Generate LLM response R on query augmented prompt\n')
|
| 237 |
prefinal_llama_with_helper_output = generate_response(modified_query)
|
|
|
|
| 238 |
return pd.Series([modified_query, prefinal_llama_with_helper_output])
|
| 239 |
|
| 240 |
|
| 241 |
def execute_pipeline(question: str):
|
| 242 |
df = pd.DataFrame({"questions": [question]})
|
| 243 |
-
print(f"
|
| 244 |
print("CUDA device name:", torch.cuda.get_device_name(0))
|
| 245 |
|
| 246 |
# queries input file
|
|
|
|
| 235 |
print('{}'.format(modified_query))
|
| 236 |
print('\nStep 7: Generate LLM response R on query augmented prompt\n')
|
| 237 |
prefinal_llama_with_helper_output = generate_response(modified_query)
|
| 238 |
+
print('{}'.format(prefinal_llama_with_helper_output))
|
| 239 |
return pd.Series([modified_query, prefinal_llama_with_helper_output])
|
| 240 |
|
| 241 |
|
| 242 |
def execute_pipeline(question: str):
|
| 243 |
df = pd.DataFrame({"questions": [question]})
|
| 244 |
+
print(f"\n\nQuestion received: {question}\n")
|
| 245 |
print("CUDA device name:", torch.cuda.get_device_name(0))
|
| 246 |
|
| 247 |
# queries input file
|