awacke1 commited on
Commit
7606437
1 Parent(s): 8890552

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -34
app.py CHANGED
@@ -229,7 +229,7 @@ def FileSidebar():
229
 
230
  # gpt
231
  try:
232
- st.write('🔍Running..')
233
  response2 = chat_with_model(user_prompt, file_contents, model_choice)
234
  filename2 = generate_filename(file_contents, choice)
235
  create_file(filename2, user_prompt, response, should_save)
@@ -242,13 +242,13 @@ def FileSidebar():
242
 
243
  if next_action=='search':
244
  file_content_area = st.text_area("File Contents:", file_contents, height=500)
245
- st.write('🔍Running..')
246
 
247
  user_prompt = file_contents
248
 
249
  # gpt
250
  try:
251
- st.write('🔍Running with GPT.')
252
  response2 = chat_with_model(user_prompt, file_contents, model_choice)
253
  filename2 = generate_filename(file_contents, choice)
254
  create_file(filename2, user_prompt, response, should_save)
@@ -356,7 +356,7 @@ def search_glossary(query):
356
 
357
  query2 = PromptPrefix + query # Add prompt preface for method step task behavior
358
  # st.write('## ' + query2)
359
- st.write('## 🔍 Running with GPT.') # -------------------------------------------------------------------------------------------------
360
  response = chat_with_model(query2)
361
  filename = generate_filename(query2 + ' --- ' + response, "md")
362
  create_file(filename, query, response, should_save)
@@ -746,7 +746,7 @@ def StreamLLMChatResponse(prompt):
746
  try:
747
  endpoint_url = API_URL
748
  hf_token = API_KEY
749
- st.write('Running client ' + endpoint_url)
750
  client = InferenceClient(endpoint_url, token=hf_token)
751
  gen_kwargs = dict(
752
  max_new_tokens=512,
@@ -1140,34 +1140,6 @@ def StreamMedChatResponse(topic):
1140
  st.write(f"Showing resources or questions related to: {topic}")
1141
 
1142
 
1143
-
1144
- def add_medical_exam_buttons():
1145
- # Medical exam terminology descriptions
1146
- descriptions = {
1147
- "White Blood Cells 🌊": "3 Q&A with emojis about types, facts, function, inputs and outputs of white blood cells 🎥",
1148
- "CT Imaging🦠": "3 Q&A with emojis on CT Imaging post surgery, how to, what to look for 💊",
1149
- "Hematoma 💉": "3 Q&A with emojis about hematoma and infection care and study including bacteria cultures and tests or labs💪",
1150
- "Post Surgery Wound Care 🍌": "3 Q&A with emojis on wound care, and good bedside manner 🩸",
1151
- "Healing and humor 💊": "3 Q&A with emojis on stories and humor about healing and caregiving 🚑",
1152
- "Psychology of bedside manner 🧬": "3 Q&A with emojis on bedside manner and how to make patients feel at ease🛠",
1153
- "CT scan 💊": "3 Q&A with analysis on infection using CT scan and packing for skin, cellulitus and fascia 🩺"
1154
- }
1155
-
1156
- # Expander for medical topics
1157
- with st.expander("Medical Licensing Exam Topics 📚", expanded=False):
1158
- st.markdown("🩺 **Important**: Variety of topics for medical licensing exams.")
1159
-
1160
- # Create buttons for each description with unique keys
1161
- for idx, (label, content) in enumerate(descriptions.items()):
1162
- button_key = f"button_{idx}"
1163
- if st.button(label, key=button_key):
1164
- st.write(f"Running {label}")
1165
- input='Create markdown outline for definition of topic ' + label + ' also short quiz with appropriate emojis and definitions for: ' + content
1166
- response=StreamLLMChatResponse(input)
1167
- filename = generate_filename(response, 'txt')
1168
- create_file(filename, input, response, should_save)
1169
-
1170
-
1171
 
1172
 
1173
  # 17. Main
@@ -1318,7 +1290,7 @@ def main():
1318
  query_params = st.query_params
1319
  #query = (query_params.get('q') or query_params.get('query') or [''])[0]
1320
  query = (query_params.get('q') or query_params.get('query') or [''])
1321
- st.markdown('# Running query: ' + query)
1322
  if query: search_glossary(query)
1323
  except:
1324
  st.markdown(' ')
 
229
 
230
  # gpt
231
  try:
232
+ #st.write('🔍Running..')
233
  response2 = chat_with_model(user_prompt, file_contents, model_choice)
234
  filename2 = generate_filename(file_contents, choice)
235
  create_file(filename2, user_prompt, response, should_save)
 
242
 
243
  if next_action=='search':
244
  file_content_area = st.text_area("File Contents:", file_contents, height=500)
245
+ #st.write('🔍Running..')
246
 
247
  user_prompt = file_contents
248
 
249
  # gpt
250
  try:
251
+ #st.write('🔍Running with GPT.')
252
  response2 = chat_with_model(user_prompt, file_contents, model_choice)
253
  filename2 = generate_filename(file_contents, choice)
254
  create_file(filename2, user_prompt, response, should_save)
 
356
 
357
  query2 = PromptPrefix + query # Add prompt preface for method step task behavior
358
  # st.write('## ' + query2)
359
+ #st.write('## 🔍 Running with GPT.') # -------------------------------------------------------------------------------------------------
360
  response = chat_with_model(query2)
361
  filename = generate_filename(query2 + ' --- ' + response, "md")
362
  create_file(filename, query, response, should_save)
 
746
  try:
747
  endpoint_url = API_URL
748
  hf_token = API_KEY
749
+ #st.write('Running client ' + endpoint_url)
750
  client = InferenceClient(endpoint_url, token=hf_token)
751
  gen_kwargs = dict(
752
  max_new_tokens=512,
 
1140
  st.write(f"Showing resources or questions related to: {topic}")
1141
 
1142
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1143
 
1144
 
1145
  # 17. Main
 
1290
  query_params = st.query_params
1291
  #query = (query_params.get('q') or query_params.get('query') or [''])[0]
1292
  query = (query_params.get('q') or query_params.get('query') or [''])
1293
+ #st.markdown('# Running query: ' + query)
1294
  if query: search_glossary(query)
1295
  except:
1296
  st.markdown(' ')