nhathuy07 commited on
Commit
cfe66d9
·
verified ·
1 Parent(s): 85e9082

Update main.py

Browse files
Files changed (1) hide show
  1. main.py +22 -13
main.py CHANGED
@@ -271,7 +271,7 @@ async def __query_ml_predict(qtype: QType, content: str, header: str, token_limi
271
  "messages": [
272
  {"role": "user", "content": prompt.gen_prompt_wh(content=content, header=header, num_qs=num_qs, lang=l)}
273
  ],
274
- "max_tokens": max(token_limit, 4096),
275
  "presence_penalty":0.3,
276
  "temperature":0.55
277
  }),
@@ -300,7 +300,7 @@ async def __query_ml_predict(qtype: QType, content: str, header: str, token_limi
300
  "messages": [
301
  {"role": "user", "content": prompt.gen_prompt_statements(content=content, header=header, num_qs=num_qs, lang=l)}
302
  ],
303
- "max_tokens": max(token_limit, 4096),
304
 
305
  }),
306
  timeout=None
@@ -328,7 +328,7 @@ async def __query_ml_predict(qtype: QType, content: str, header: str, token_limi
328
  "messages": [
329
  {"role": "user", "content": prompt.gen_prompt_statements_false(content=_r_content, lang=l)}
330
  ],
331
- "max_tokens": max(token_limit, 4096),
332
 
333
  }),
334
  timeout=None
@@ -532,16 +532,25 @@ async def generate_questions(request):
532
  if (o['style'] == QType.WH):
533
 
534
  raw_qa_list = []
535
- raw_segmented: list[str] = list(filter(lambda x: (len(x)>0), o['content'].split("\n\n")))[1:]
536
- print(raw_segmented)
537
- for i in range(len(raw_segmented)):
538
- if (len(raw_segmented[i]) and raw_segmented[i].count('\n') < 5):
539
- raw_segmented[i] += f'\n{raw_segmented[i+1]}'
540
- raw_segmented[i+1] = ""
541
-
542
- print(raw_segmented)
543
-
544
- __parsed_outputs.extend(await parse_wh_question(raw_segmented, pgph_i))
 
 
 
 
 
 
 
 
 
545
 
546
  elif (o['style'] == QType.STMT):
547
  print(o['content'])
 
271
  "messages": [
272
  {"role": "user", "content": prompt.gen_prompt_wh(content=content, header=header, num_qs=num_qs, lang=l)}
273
  ],
274
+ "max_tokens": 4096,
275
  "presence_penalty":0.3,
276
  "temperature":0.55
277
  }),
 
300
  "messages": [
301
  {"role": "user", "content": prompt.gen_prompt_statements(content=content, header=header, num_qs=num_qs, lang=l)}
302
  ],
303
+ "max_tokens": 4096,
304
 
305
  }),
306
  timeout=None
 
328
  "messages": [
329
  {"role": "user", "content": prompt.gen_prompt_statements_false(content=_r_content, lang=l)}
330
  ],
331
+ "max_tokens": 4096,
332
 
333
  }),
334
  timeout=None
 
532
  if (o['style'] == QType.WH):
533
 
534
  raw_qa_list = []
535
+ # raw_segmented: list[str] = list(filter(lambda x: (len(x)>0), o['content'].split("\n\n")))[1:]
536
+
537
+ # for i in range(len(raw_segmented)):
538
+ # if (len(raw_segmented[i]) and raw_segmented[i].count('\n') < 5):
539
+ # raw_segmented[i] += f'\n{raw_segmented[i+1]}'
540
+ # raw_segmented[i+1] = ""
541
+
542
+ # print(raw_segmented)
543
+ seg_index = 0
544
+ seg_index_map = ['Q.', 'A.', 'B.', 'C.', 'D.', '{']
545
+ raw_segmented = []
546
+ for seg in o['content'].split('\n'):
547
+ if seg.startswith(seg_index_map[seg_index]):
548
+ raw_segmented.append(seg)
549
+ seg_index+=1
550
+ if seg_index == 5:
551
+ __parsed_outputs.extend(await parse_wh_question(raw_segmented, pgph_i))
552
+ break
553
+ seg_index = 0
554
 
555
  elif (o['style'] == QType.STMT):
556
  print(o['content'])