tamas.kiss commited on
Commit
d56b068
1 Parent(s): 245d66c

Update logs

Browse files
Files changed (1) hide show
  1. app.py +10 -2
app.py CHANGED
@@ -258,8 +258,6 @@ def text_to_text_generation(verbose, prompt):
258
  )
259
  response = response[len(is_kubectl_prompt) :]
260
 
261
- print(f'{" Query Start ":-^40}')
262
- print("Classified as: " + response)
263
 
264
  response_num = 0 if "0" in response else (1 if "1" in response else 2)
265
 
@@ -310,11 +308,17 @@ def text_to_text_generation(verbose, prompt):
310
  return (prompt, answer[start:end].strip())
311
 
312
  modes = ["Kubectl command", "Kubernetes definition", "Normal"]
 
 
 
 
313
  modes[response_num] = f"**{modes[response_num]}**"
314
  modes = " / ".join(modes)
315
 
 
316
  if response_num == 2:
317
  prompt = create_generation_prompt(response_num, prompt, False)
 
318
  original, new = generate_batch(prompt)[0]
319
  prompt, response = cleanup(original, new)
320
  if verbose:
@@ -324,6 +328,7 @@ def text_to_text_generation(verbose, prompt):
324
 
325
  if response_num == 0:
326
  prompt = create_generation_prompt(response_num, prompt, False)
 
327
  original, new = generate_batch(prompt)[0]
328
  prompt, response = cleanup(original, new)
329
  model_response = new[len(original):].strip()
@@ -341,8 +346,11 @@ def text_to_text_generation(verbose, prompt):
341
  return f"{modes}\n\n" f"# Answer:\n" f"```bash\n{str_to_md(response)}\n```\n"
342
 
343
  res_prompt = create_generation_prompt(response_num, prompt, False)
 
344
  res_semantic_search_prompt = create_generation_prompt(response_num, prompt, "semantic_search")
 
345
  res_google_search_prompt = create_generation_prompt(response_num, prompt, "google_search")
 
346
 
347
  gen_normal, gen_semantic_search, gen_google_search = generate_batch(
348
  res_prompt, res_semantic_search_prompt, res_google_search_prompt
 
258
  )
259
  response = response[len(is_kubectl_prompt) :]
260
 
 
 
261
 
262
  response_num = 0 if "0" in response else (1 if "1" in response else 2)
263
 
 
308
  return (prompt, answer[start:end].strip())
309
 
310
  modes = ["Kubectl command", "Kubernetes definition", "Normal"]
311
+
312
+ print(f'{" Query Start ":-^40}')
313
+ print("Classified as: " + modes[response_num])
314
+
315
  modes[response_num] = f"**{modes[response_num]}**"
316
  modes = " / ".join(modes)
317
 
318
+
319
  if response_num == 2:
320
  prompt = create_generation_prompt(response_num, prompt, False)
321
+ print('Prompt given to model:\n' + prompt + '\n')
322
  original, new = generate_batch(prompt)[0]
323
  prompt, response = cleanup(original, new)
324
  if verbose:
 
328
 
329
  if response_num == 0:
330
  prompt = create_generation_prompt(response_num, prompt, False)
331
+ print('Prompt given to model:\n' + prompt + '\n')
332
  original, new = generate_batch(prompt)[0]
333
  prompt, response = cleanup(original, new)
334
  model_response = new[len(original):].strip()
 
346
  return f"{modes}\n\n" f"# Answer:\n" f"```bash\n{str_to_md(response)}\n```\n"
347
 
348
  res_prompt = create_generation_prompt(response_num, prompt, False)
349
+ print(f'Prompt given to finetuned model:\n{res_google_search_prompt}\n')
350
  res_semantic_search_prompt = create_generation_prompt(response_num, prompt, "semantic_search")
351
+ print(f'Prompt given to model with RAG:\n{res_semantic_search_prompt}\n')
352
  res_google_search_prompt = create_generation_prompt(response_num, prompt, "google_search")
353
+ print(f'Prompt given to model with Google search:\n{res_google_search_prompt}\n')
354
 
355
  gen_normal, gen_semantic_search, gen_google_search = generate_batch(
356
  res_prompt, res_semantic_search_prompt, res_google_search_prompt