|
{ |
|
"config_general": { |
|
"model_name": "yi-vision", |
|
"model_dtype": "float16", |
|
"model_size": 0 |
|
}, |
|
"results": { |
|
"ChartQA": { |
|
"acc": 78.88, |
|
"acc_stderr": 0, |
|
"accuracy": 78.88, |
|
"human_test": { |
|
"total": 1250, |
|
"correct": 811, |
|
"accuracy": 64.88 |
|
}, |
|
"augmented_test": { |
|
"total": 1250, |
|
"correct": 1161, |
|
"accuracy": 92.88 |
|
} |
|
}, |
|
"CMMMU": { |
|
"acc": 41.22, |
|
"acc_stderr": 0, |
|
"\u5546\u4e1a": { |
|
"num": 126, |
|
"correct": 30, |
|
"accuracy": 23.81 |
|
}, |
|
"\u79d1\u5b66": { |
|
"num": 204, |
|
"correct": 78, |
|
"accuracy": 38.24 |
|
}, |
|
"overall": { |
|
"num": 900, |
|
"correct": 371, |
|
"accuracy": 41.22 |
|
}, |
|
"accuracy": 41.22, |
|
"\u5065\u5eb7\u4e0e\u533b\u5b66": { |
|
"num": 153, |
|
"correct": 77, |
|
"accuracy": 50.33 |
|
}, |
|
"\u6280\u672f\u4e0e\u5de5\u7a0b": { |
|
"num": 244, |
|
"correct": 89, |
|
"accuracy": 36.48 |
|
}, |
|
"\u827a\u672f\u4e0e\u8bbe\u8ba1": { |
|
"num": 88, |
|
"correct": 56, |
|
"accuracy": 63.64 |
|
}, |
|
"\u4eba\u6587\u793e\u4f1a\u79d1\u5b66": { |
|
"num": 85, |
|
"correct": 41, |
|
"accuracy": 48.24 |
|
} |
|
}, |
|
"CMMU": { |
|
"acc": 22.97, |
|
"acc_stderr": 0, |
|
"val": { |
|
"multiple-choice": { |
|
"hard": { |
|
"total": 150, |
|
"correct": 30, |
|
"accuracy": 20.0 |
|
}, |
|
"normal": { |
|
"total": 1205, |
|
"correct": 361, |
|
"accuracy": 29.96 |
|
} |
|
}, |
|
"fill-in-the-blank": { |
|
"hard": { |
|
"total": 300, |
|
"correct": 44, |
|
"accuracy": 14.67 |
|
}, |
|
"normal": { |
|
"total": 507, |
|
"correct": 77, |
|
"accuracy": 15.19 |
|
} |
|
}, |
|
"multiple-response": { |
|
"hard": { |
|
"total": 94, |
|
"correct": 9, |
|
"accuracy": 9.57 |
|
}, |
|
"normal": { |
|
"total": 33, |
|
"correct": 6, |
|
"accuracy": 18.18 |
|
} |
|
} |
|
}, |
|
"test": { |
|
"multiple-choice": { |
|
"hard": { |
|
"total": 150, |
|
"correct": 25, |
|
"accuracy": 16.67 |
|
}, |
|
"normal": { |
|
"total": 1205, |
|
"correct": 380, |
|
"accuracy": 31.54 |
|
} |
|
}, |
|
"fill-in-the-blank": { |
|
"hard": { |
|
"total": 296, |
|
"correct": 37, |
|
"accuracy": 12.5 |
|
}, |
|
"normal": { |
|
"total": 529, |
|
"correct": 76, |
|
"accuracy": 14.37 |
|
} |
|
}, |
|
"multiple-response": { |
|
"hard": { |
|
"total": 95, |
|
"correct": 8, |
|
"accuracy": 8.42 |
|
}, |
|
"normal": { |
|
"total": 32, |
|
"correct": 4, |
|
"accuracy": 12.5 |
|
} |
|
} |
|
}, |
|
"val-overall": { |
|
"total": 2289, |
|
"correct": 527, |
|
"accuracy": 23.02, |
|
"bias_rate": 7.13 |
|
}, |
|
"test-overall": { |
|
"total": 2307, |
|
"correct": 530, |
|
"accuracy": 22.97, |
|
"bias_rate": 5.58 |
|
} |
|
}, |
|
"MMMU": { |
|
"acc": 46.33, |
|
"acc_stderr": 0, |
|
"accuracy": 46.33, |
|
"subject_score": { |
|
"Art": 71.67, |
|
"Math": 46.67, |
|
"Basic": 53.33, |
|
"Music": 30.0, |
|
"Design": 80.0, |
|
"Energy": 30.0, |
|
"Manage": 40.0, |
|
"Public": 46.67, |
|
"Biology": 46.67, |
|
"Finance": 36.67, |
|
"History": 63.33, |
|
"Physics": 40.0, |
|
"Clinical": 53.33, |
|
"Computer": 40.0, |
|
"Pharmacy": 30.0, |
|
"Chemistry": 23.33, |
|
"Economics": 60.0, |
|
"Geography": 60.0, |
|
"Marketing": 53.33, |
|
"Materials": 20.0, |
|
"Sociology": 56.67, |
|
"Accounting": 36.67, |
|
"Literature": 83.33, |
|
"Mechanical": 23.33, |
|
"Psychology": 46.67, |
|
"Agriculture": 53.33, |
|
"Diagnostics": 43.33, |
|
"Electronics": 23.33, |
|
"Architecture": 26.67 |
|
}, |
|
"difficulty_score": { |
|
"Easy": 54.92, |
|
"Hard": 30.39, |
|
"Medium": 47.17 |
|
} |
|
}, |
|
"MMMU_Pro_standard": { |
|
"acc": 29.84, |
|
"acc_stderr": 0, |
|
"accuracy": 29.84, |
|
"reject_info": { |
|
"reject_rate": 0.06, |
|
"reject_number": 1, |
|
"total_question": 1730 |
|
}, |
|
"subject_score": { |
|
"Art": 47.17, |
|
"Math": 21.67, |
|
"Music": 20.0, |
|
"Design": 47.46, |
|
"Manage": 24.0, |
|
"Biology": 32.2, |
|
"Finance": 25.0, |
|
"History": 42.86, |
|
"Physics": 23.33, |
|
"Pharmacy": 31.58, |
|
"Chemistry": 21.67, |
|
"Economics": 23.73, |
|
"Geography": 36.54, |
|
"Marketing": 33.9, |
|
"Materials": 13.33, |
|
"Sociology": 48.15, |
|
"Accounting": 24.14, |
|
"Art_Theory": 60.0, |
|
"Literature": 73.08, |
|
"Psychology": 28.33, |
|
"Agriculture": 25.0, |
|
"Electronics": 21.67, |
|
"Public_Health": 24.14, |
|
"Computer_Science": 30.0, |
|
"Energy_and_Power": 12.07, |
|
"Clinical_Medicine": 27.12, |
|
"Basic_Medical_Science": 34.62, |
|
"Mechanical_Engineering": 20.34, |
|
"Architecture_and_Engineering": 10.0, |
|
"Diagnostics_and_Laboratory_Medicine": 25.0 |
|
}, |
|
"difficulty_score": { |
|
"Easy": 43.75, |
|
"Hard": 17.96, |
|
"Medium": 26.62 |
|
} |
|
}, |
|
"MMMU_Pro_vision":{ |
|
"acc":53.06 |
|
}, |
|
"OCRBench": { |
|
"acc": 69.369, |
|
"acc_stderr": 0, |
|
"accuracy": 69.369, |
|
"final_score": [ |
|
693, |
|
999 |
|
], |
|
"reject_info": { |
|
"reject_rate": 0.1, |
|
"reject_number": 1, |
|
"total_question": 1000 |
|
}, |
|
"Doc-oriented VQA": [ |
|
146, |
|
200 |
|
], |
|
"Scene Text-centric VQA": [ |
|
174, |
|
200 |
|
], |
|
"Handwriting Recognition": [ |
|
28, |
|
50 |
|
], |
|
"Digit String Recognition": [ |
|
21, |
|
50 |
|
], |
|
"Regular Text Recognition": [ |
|
48, |
|
50 |
|
], |
|
"Artistic Text Recognition": [ |
|
42, |
|
50 |
|
], |
|
"Irregular Text Recognition": [ |
|
40, |
|
50 |
|
], |
|
"Key Information Extraction": [ |
|
154, |
|
200 |
|
], |
|
"Non-Semantic Text Recognition": [ |
|
38, |
|
50 |
|
], |
|
"Handwritten Mathematical Expression Recognition": [ |
|
2, |
|
99 |
|
] |
|
}, |
|
"MathVision": { |
|
"acc": 14.44, |
|
"acc_stderr": 0, |
|
"accuracy": 14.44 |
|
}, |
|
"CII-Bench": { |
|
"acc": 45.23, |
|
"acc_stderr": 0, |
|
"accuracy": 45.23, |
|
"domain_score": { |
|
"Art": 49.26, |
|
"CTC": 48.15, |
|
"Env.": 53.7, |
|
"Life": 37.23, |
|
"Society": 45.95, |
|
"Politics": 58.33 |
|
}, |
|
"emotion_score": { |
|
"Neutral": 48.5, |
|
"Negative": 42.64, |
|
"Positive": 44.44 |
|
} |
|
}, |
|
"Blink":{ |
|
"acc":0 |
|
}, |
|
}, |
|
"versions": {}, |
|
"config_tasks": {}, |
|
"summary_tasks": {}, |
|
"summary_general": {} |
|
} |