Spaces:
Running
Running
Update main.py
Browse files
main.py
CHANGED
@@ -271,7 +271,7 @@ async def __query_ml_predict(qtype: QType, content: str, header: str, token_limi
|
|
271 |
"messages": [
|
272 |
{"role": "user", "content": prompt.gen_prompt_wh(content=content, header=header, num_qs=num_qs, lang=l)}
|
273 |
],
|
274 |
-
"max_tokens":
|
275 |
"presence_penalty":0.3,
|
276 |
"temperature":0.55
|
277 |
}),
|
@@ -300,7 +300,7 @@ async def __query_ml_predict(qtype: QType, content: str, header: str, token_limi
|
|
300 |
"messages": [
|
301 |
{"role": "user", "content": prompt.gen_prompt_statements(content=content, header=header, num_qs=num_qs, lang=l)}
|
302 |
],
|
303 |
-
"max_tokens":
|
304 |
|
305 |
}),
|
306 |
timeout=None
|
@@ -328,7 +328,7 @@ async def __query_ml_predict(qtype: QType, content: str, header: str, token_limi
|
|
328 |
"messages": [
|
329 |
{"role": "user", "content": prompt.gen_prompt_statements_false(content=_r_content, lang=l)}
|
330 |
],
|
331 |
-
"max_tokens":
|
332 |
|
333 |
}),
|
334 |
timeout=None
|
@@ -532,16 +532,25 @@ async def generate_questions(request):
|
|
532 |
if (o['style'] == QType.WH):
|
533 |
|
534 |
raw_qa_list = []
|
535 |
-
raw_segmented: list[str] = list(filter(lambda x: (len(x)>0), o['content'].split("\n\n")))[1:]
|
536 |
-
|
537 |
-
for i in range(len(raw_segmented)):
|
538 |
-
|
539 |
-
|
540 |
-
|
541 |
-
|
542 |
-
print(raw_segmented)
|
543 |
-
|
544 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
545 |
|
546 |
elif (o['style'] == QType.STMT):
|
547 |
print(o['content'])
|
|
|
271 |
"messages": [
|
272 |
{"role": "user", "content": prompt.gen_prompt_wh(content=content, header=header, num_qs=num_qs, lang=l)}
|
273 |
],
|
274 |
+
"max_tokens": 4096,
|
275 |
"presence_penalty":0.3,
|
276 |
"temperature":0.55
|
277 |
}),
|
|
|
300 |
"messages": [
|
301 |
{"role": "user", "content": prompt.gen_prompt_statements(content=content, header=header, num_qs=num_qs, lang=l)}
|
302 |
],
|
303 |
+
"max_tokens": 4096,
|
304 |
|
305 |
}),
|
306 |
timeout=None
|
|
|
328 |
"messages": [
|
329 |
{"role": "user", "content": prompt.gen_prompt_statements_false(content=_r_content, lang=l)}
|
330 |
],
|
331 |
+
"max_tokens": 4096,
|
332 |
|
333 |
}),
|
334 |
timeout=None
|
|
|
532 |
if (o['style'] == QType.WH):
|
533 |
|
534 |
raw_qa_list = []
|
535 |
+
# raw_segmented: list[str] = list(filter(lambda x: (len(x)>0), o['content'].split("\n\n")))[1:]
|
536 |
+
|
537 |
+
# for i in range(len(raw_segmented)):
|
538 |
+
# if (len(raw_segmented[i]) and raw_segmented[i].count('\n') < 5):
|
539 |
+
# raw_segmented[i] += f'\n{raw_segmented[i+1]}'
|
540 |
+
# raw_segmented[i+1] = ""
|
541 |
+
|
542 |
+
# print(raw_segmented)
|
543 |
+
seg_index = 0
|
544 |
+
seg_index_map = ['Q.', 'A.', 'B.', 'C.', 'D.', '{']
|
545 |
+
raw_segmented = []
|
546 |
+
for seg in o['content'].split('\n'):
|
547 |
+
if seg.startswith(seg_index_map[seg_index]):
|
548 |
+
raw_segmented.append(seg)
|
549 |
+
seg_index+=1
|
550 |
+
if seg_index == 5:
|
551 |
+
__parsed_outputs.extend(await parse_wh_question(raw_segmented, pgph_i))
|
552 |
+
break
|
553 |
+
seg_index = 0
|
554 |
|
555 |
elif (o['style'] == QType.STMT):
|
556 |
print(o['content'])
|