aatu18 commited on
Commit
00d709d
·
verified ·
1 Parent(s): 450cf49

add output for step 7

Browse files
Files changed (1) hide show
  1. app.py +2 -1
app.py CHANGED
@@ -235,12 +235,13 @@ def get_prefinal_response(row):
235
  print('{}'.format(modified_query))
236
  print('\nStep 7: Generate LLM response R on query augmented prompt\n')
237
  prefinal_llama_with_helper_output = generate_response(modified_query)
 
238
  return pd.Series([modified_query, prefinal_llama_with_helper_output])
239
 
240
 
241
  def execute_pipeline(question: str):
242
  df = pd.DataFrame({"questions": [question]})
243
- print(f"Question received: {question}")
244
  print("CUDA device name:", torch.cuda.get_device_name(0))
245
 
246
  # queries input file
 
235
  print('{}'.format(modified_query))
236
  print('\nStep 7: Generate LLM response R on query augmented prompt\n')
237
  prefinal_llama_with_helper_output = generate_response(modified_query)
238
+ print('{}'.format(prefinal_llama_with_helper_output))
239
  return pd.Series([modified_query, prefinal_llama_with_helper_output])
240
 
241
 
242
  def execute_pipeline(question: str):
243
  df = pd.DataFrame({"questions": [question]})
244
+ print(f"\n\nQuestion received: {question}\n")
245
  print("CUDA device name:", torch.cuda.get_device_name(0))
246
 
247
  # queries input file