awacke1 commited on
Commit
ce2a709
1 Parent(s): aacfca5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -113,7 +113,7 @@ def StreamLLMChatResponse(prompt):
113
  res_box = st.empty()
114
  collected_chunks=[]
115
  collected_messages=[]
116
- allresults=''
117
  for r in stream:
118
  if r.token.special:
119
  continue
@@ -127,12 +127,12 @@ def StreamLLMChatResponse(prompt):
127
  if len(r.token.text) > 0:
128
  result="".join(report).strip()
129
  res_box.markdown(f'*{result}*')
130
- allresults=allresults+result
131
  except:
132
  st.write('Stream llm issue')
133
 
134
- st.write(allresults)
135
- return allresults
136
  except:
137
  st.write('DromeLlama is asleep. Starting up now on A10 - please give 5 minutes then retry as KEDA scales up from zero to activate running container(s).')
138
 
 
113
  res_box = st.empty()
114
  collected_chunks=[]
115
  collected_messages=[]
116
+ #allresults=''
117
  for r in stream:
118
  if r.token.special:
119
  continue
 
127
  if len(r.token.text) > 0:
128
  result="".join(report).strip()
129
  res_box.markdown(f'*{result}*')
130
+ #allresults=allresults+result
131
  except:
132
  st.write('Stream llm issue')
133
 
134
+ #st.write(allresults)
135
+ #return allresults
136
  except:
137
  st.write('DromeLlama is asleep. Starting up now on A10 - please give 5 minutes then retry as KEDA scales up from zero to activate running container(s).')
138