File size: 167,502 Bytes
a5038f9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 306 307 308 309 310 311 312 313 314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 475 476 477 478 479 480 481 482 483 484 485 486 487 488 489 490 491 492 493 494 495 496 497 498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 570 571 572 573 574 575 576 577 578 579 580 581 582 |
/mnt/petrelfs/wangweiyun/miniconda3/envs/internvl_eval2/lib/python3.10/site-packages/bitsandbytes/cextension.py:34: UserWarning: The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable.
warn("The installed version of bitsandbytes was compiled without GPU support. "
/mnt/petrelfs/wangweiyun/miniconda3/envs/internvl_eval2/lib/python3.10/site-packages/bitsandbytes/libbitsandbytes_cpu.so: undefined symbol: cadam32bit_grad_fp32
model path is /mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B
12/05 03:39:44 - OpenCompass - WARNING - No previous results to reuse!
12/05 03:39:44 - OpenCompass - INFO - Reusing experiements from 20241205_033944
12/05 03:39:44 - OpenCompass - INFO - Current exp folder: /mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B/20241205_033944
12/05 03:39:48 - OpenCompass - INFO - Partitioned into 256 tasks.
[ ] 0/256, elapsed: 0s, ETA:use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26604 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99913_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13280 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100181_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24526 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100186_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14432 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100185_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30532 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100173_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24834 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99968_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25967 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100187_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24407 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100182_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26493 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99942_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=21062 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99920_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31494 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100177_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27535 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99966_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16310 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100171_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=21774 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99898_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12469 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100166_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14638 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99929_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16693 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99888_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27499 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99952_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16708 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100060_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27588 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99988_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25288 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99889_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31890 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100174_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26297 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100180_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16397 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100157_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12164 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100168_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26540 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100176_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29112 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100179_params.py
command torchrun --master_port=23949 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99887_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=18811 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99994_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13418 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100169_params.py
use_backend False use_backend{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29222 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100032_params.py
command torchrun --master_port=18950 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99947_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=21842 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100162_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30688 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99870_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17817 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100149_params.py
use_backenduse_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16554 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100139_params.py
False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=18486 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99855_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=18748 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100178_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28115 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100184_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15313 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100158_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28444 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100151_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29823 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99914_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28431 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99932_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30982 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99863_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22312 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99901_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26987 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100175_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19899 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99983_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19490 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100167_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=23066 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100133_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30647 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100135_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13344 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99979_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20192 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99893_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19336 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100160_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22730 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100132_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14349 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100122_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22606 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100141_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31476 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99916_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12584 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99921_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17514 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100142_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22902 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100170_params.py
command torchrun --master_port=17978 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100172_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17883 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99948_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25101 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99909_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12829 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100119_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14215 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100153_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24184 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99939_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24341 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100147_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12512 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99970_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26047 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100134_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12077 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99874_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14692 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100188_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27755 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100116_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29389 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99877_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=18251 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100127_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30001 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99895_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29486 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99886_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24196 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100138_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17277 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99982_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30313 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100155_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25681 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100164_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20526 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99977_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14619 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100111_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26666 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100144_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20493 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99866_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28853 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99962_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22252 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100101_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24525 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100108_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=23600 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100114_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15879 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100159_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12811 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100107_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17538 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100125_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22218 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100183_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26764 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100150_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28394 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100120_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15424 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100078_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19811 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100148_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30170 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99987_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19203 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99900_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12951 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99928_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29541 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99910_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=21271 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99926_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=23626 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99919_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17356 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99931_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16269 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100140_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24375 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100115_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20257 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99915_params.py
use_backend False use_backend False{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15094 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99985_params.py
command torchrun --master_port=15353 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100082_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17610 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99878_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13371 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99986_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29090 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100123_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=21099 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99892_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29294 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100145_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=21659 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100165_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16912 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100152_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13163 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99865_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16231 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100124_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12932 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100083_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14734 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99958_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31341 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100073_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14328 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99876_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20416 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100113_params.py
use_backend False use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command commandtorchrun --master_port=17135 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99879_params.py
torchrun --master_port=18677 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99924_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12094 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99912_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19355 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100112_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15928 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100104_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12862 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100106_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27787 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99864_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13000 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100090_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28745 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100063_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25264 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100126_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16067 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100146_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27299 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100096_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16906 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100156_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19369 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100110_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26611 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99961_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16376 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99963_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27537 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100088_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=21108 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99941_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17468 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100102_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20674 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99957_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29071 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100163_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15347 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100077_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24037 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100131_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12977 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100072_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13057 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100109_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13267 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99959_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15163 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99972_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25976 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100079_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27184 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99944_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31203 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100089_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27583 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99976_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30842 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100118_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25586 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100121_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19830 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100161_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31081 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100095_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29153 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100075_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=23362 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100081_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29435 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99953_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12824 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100093_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=18395 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99884_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24334 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99881_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20392 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99896_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19381 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99984_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16250 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100080_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20045 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100098_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12230 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99908_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=21004 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100117_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30647 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99872_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19046 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99989_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19156 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100062_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28238 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99956_params.py
use_backenduse_backend False False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16891 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100094_params.py
command torchrun --master_port=15867 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99930_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31662 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100128_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28884 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100071_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30497 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100092_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17261 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99954_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25699 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100129_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30047 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99981_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13601 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100066_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}command torchrun --master_port=26219 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100130_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19258 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100154_params.py
command torchrun --master_port=12233 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99894_params.py
use_backenduse_backend
False False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22309 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100084_params.py
command torchrun --master_port=23255 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100074_params.py
command torchrun --master_port=12847 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100086_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22258 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99868_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=18497 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99967_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29005 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100076_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=23792 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100091_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16578 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99885_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17748 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99991_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30847 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100068_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24524 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99990_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22131 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100085_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15395 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99995_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12427 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100100_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15801 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100099_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31270 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99940_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29268 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99978_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16793 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100065_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17810 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99993_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15272 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99992_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13425 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99936_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=19139 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100069_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26643 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99973_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27017 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100097_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=20868 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100103_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25388 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99883_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29743 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99975_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31817 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99911_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13689 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99933_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26617 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100070_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22327 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99918_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12446 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99938_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28013 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99969_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13858 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99974_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25635 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99923_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17105 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99880_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31600 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99927_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27439 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100105_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=24107 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99859_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28968 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99903_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=15510 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100064_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=18400 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99945_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27945 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99875_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31191 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99899_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28700 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99971_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22613 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99907_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30183 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100087_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=14671 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99935_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25091 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99922_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22640 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99917_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22182 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99960_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17209 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99867_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13371 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99858_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=27849 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99856_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=23151 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99897_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22044 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99965_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28553 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99950_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=12281 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99980_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16734 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99860_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=31318 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99861_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16303 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99882_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=13460 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99873_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26241 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99869_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=23737 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99857_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=28106 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99906_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29863 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99854_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=29951 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99853_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16872 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99902_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30472 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99964_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=25834 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99871_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=17323 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99949_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=30716 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100187_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=16512 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99888_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=22260 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99966_params.py
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
command torchrun --master_port=26044 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99898_params.py
[ ] 1/256, 0.0 task/s, elapsed: 205s, ETA: 52230s
[ ] 2/256, 0.0 task/s, elapsed: 209s, ETA: 26539s
[ ] 3/256, 0.0 task/s, elapsed: 219s, ETA: 18510s
[ ] 4/256, 0.0 task/s, elapsed: 223s, ETA: 14078s
[ ] 5/256, 0.0 task/s, elapsed: 225s, ETA: 11290s
[ ] 6/256, 0.0 task/s, elapsed: 226s, ETA: 9431s
[ ] 7/256, 0.0 task/s, elapsed: 231s, ETA: 8229s
[ ] 8/256, 0.0 task/s, elapsed: 235s, ETA: 7277s
[> ] 9/256, 0.0 task/s, elapsed: 237s, ETA: 6506s
[> ] 10/256, 0.0 task/s, elapsed: 240s, ETA: 5909s
[> ] 11/256, 0.0 task/s, elapsed: 241s, ETA: 5366s
[> ] 12/256, 0.0 task/s, elapsed: 242s, ETA: 4913s
[> ] 13/256, 0.1 task/s, elapsed: 244s, ETA: 4569s
[> ] 14/256, 0.1 task/s, elapsed: 246s, ETA: 4259s
[> ] 15/256, 0.1 task/s, elapsed: 250s, ETA: 4013s
[> ] 16/256, 0.1 task/s, elapsed: 251s, ETA: 3760s
[> ] 17/256, 0.1 task/s, elapsed: 251s, ETA: 3534s
[>> ] 18/256, 0.1 task/s, elapsed: 254s, ETA: 3364s
[>> ] 19/256, 0.1 task/s, elapsed: 256s, ETA: 3199s
[>> ] 20/256, 0.1 task/s, elapsed: 259s, ETA: 3056s
[>> ] 21/256, 0.1 task/s, elapsed: 261s, ETA: 2917s
[>> ] 22/256, 0.1 task/s, elapsed: 264s, ETA: 2804s
[>> ] 23/256, 0.1 task/s, elapsed: 264s, ETA: 2676s
[>> ] 24/256, 0.1 task/s, elapsed: 264s, ETA: 2554s
[>> ] 25/256, 0.1 task/s, elapsed: 266s, ETA: 2461s
[>>> ] 26/256, 0.1 task/s, elapsed: 271s, ETA: 2393s
[>>> ] 27/256, 0.1 task/s, elapsed: 276s, ETA: 2337s
[>>> ] 28/256, 0.1 task/s, elapsed: 279s, ETA: 2273s
[>>> ] 29/256, 0.1 task/s, elapsed: 281s, ETA: 2203s
[>>> ] 30/256, 0.1 task/s, elapsed: 282s, ETA: 2123s
[>>> ] 31/256, 0.1 task/s, elapsed: 282s, ETA: 2049s
[>>> ] 32/256, 0.1 task/s, elapsed: 283s, ETA: 1978s
[>>> ] 33/256, 0.1 task/s, elapsed: 284s, ETA: 1916s
[>>> ] 34/256, 0.1 task/s, elapsed: 284s, ETA: 1856s
[>>>> ] 35/256, 0.1 task/s, elapsed: 286s, ETA: 1806s
[>>>> ] 36/256, 0.1 task/s, elapsed: 288s, ETA: 1758s
[>>>> ] 37/256, 0.1 task/s, elapsed: 288s, ETA: 1706s
[>>>> ] 38/256, 0.1 task/s, elapsed: 290s, ETA: 1666s
[>>>> ] 39/256, 0.1 task/s, elapsed: 292s, ETA: 1623s
[>>>> ] 40/256, 0.1 task/s, elapsed: 292s, ETA: 1578s
[>>>> ] 41/256, 0.1 task/s, elapsed: 292s, ETA: 1532s
[>>>> ] 42/256, 0.1 task/s, elapsed: 294s, ETA: 1497s
[>>>>> ] 43/256, 0.1 task/s, elapsed: 294s, ETA: 1456s
[>>>>> ] 44/256, 0.1 task/s, elapsed: 295s, ETA: 1421s
[>>>>> ] 45/256, 0.2 task/s, elapsed: 296s, ETA: 1386s
[>>>>> ] 46/256, 0.2 task/s, elapsed: 296s, ETA: 1353s
[>>>>> ] 47/256, 0.2 task/s, elapsed: 296s, ETA: 1318s
[>>>>> ] 48/256, 0.2 task/s, elapsed: 297s, ETA: 1285s
[>>>>> ] 49/256, 0.2 task/s, elapsed: 298s, ETA: 1259s
[>>>>> ] 50/256, 0.2 task/s, elapsed: 299s, ETA: 1230s
[>>>>> ] 51/256, 0.2 task/s, elapsed: 299s, ETA: 1202s
[>>>>>> ] 52/256, 0.2 task/s, elapsed: 299s, ETA: 1173s
[>>>>>> ] 53/256, 0.2 task/s, elapsed: 300s, ETA: 1150s
[>>>>>> ] 54/256, 0.2 task/s, elapsed: 301s, ETA: 1127s
[>>>>>> ] 55/256, 0.2 task/s, elapsed: 302s, ETA: 1103s
[>>>>>> ] 56/256, 0.2 task/s, elapsed: 303s, ETA: 1083s
[>>>>>> ] 57/256, 0.2 task/s, elapsed: 304s, ETA: 1060s
[>>>>>> ] 58/256, 0.2 task/s, elapsed: 304s, ETA: 1037s
[>>>>>> ] 59/256, 0.2 task/s, elapsed: 304s, ETA: 1014s
[>>>>>>> ] 60/256, 0.2 task/s, elapsed: 304s, ETA: 994s
[>>>>>>> ] 61/256, 0.2 task/s, elapsed: 305s, ETA: 976s
[>>>>>>> ] 62/256, 0.2 task/s, elapsed: 305s, ETA: 955s
[>>>>>>> ] 63/256, 0.2 task/s, elapsed: 306s, ETA: 939s
[>>>>>>> ] 64/256, 0.2 task/s, elapsed: 309s, ETA: 927s
[>>>>>>> ] 65/256, 0.2 task/s, elapsed: 311s, ETA: 915s
[>>>>>>> ] 66/256, 0.2 task/s, elapsed: 312s, ETA: 897s
[>>>>>>> ] 67/256, 0.2 task/s, elapsed: 312s, ETA: 880s
[>>>>>>> ] 68/256, 0.2 task/s, elapsed: 315s, ETA: 870s
[>>>>>>>> ] 69/256, 0.2 task/s, elapsed: 315s, ETA: 853s
[>>>>>>>> ] 70/256, 0.2 task/s, elapsed: 315s, ETA: 837s
[>>>>>>>> ] 71/256, 0.2 task/s, elapsed: 315s, ETA: 821s
[>>>>>>>> ] 72/256, 0.2 task/s, elapsed: 316s, ETA: 806s
[>>>>>>>> ] 73/256, 0.2 task/s, elapsed: 317s, ETA: 795s
[>>>>>>>> ] 74/256, 0.2 task/s, elapsed: 318s, ETA: 782s
[>>>>>>>> ] 75/256, 0.2 task/s, elapsed: 319s, ETA: 770s
[>>>>>>>> ] 76/256, 0.2 task/s, elapsed: 319s, ETA: 757s
[>>>>>>>>> ] 77/256, 0.2 task/s, elapsed: 320s, ETA: 743s
[>>>>>>>>> ] 78/256, 0.2 task/s, elapsed: 320s, ETA: 730s
[>>>>>>>>> ] 79/256, 0.2 task/s, elapsed: 322s, ETA: 721s
[>>>>>>>>> ] 80/256, 0.2 task/s, elapsed: 323s, ETA: 711s
[>>>>>>>>> ] 81/256, 0.3 task/s, elapsed: 324s, ETA: 699s
[>>>>>>>>> ] 82/256, 0.3 task/s, elapsed: 324s, ETA: 688s
[>>>>>>>>> ] 83/256, 0.3 task/s, elapsed: 324s, ETA: 675s
[>>>>>>>>> ] 84/256, 0.3 task/s, elapsed: 326s, ETA: 667s
[>>>>>>>>> ] 85/256, 0.3 task/s, elapsed: 326s, ETA: 657s
[>>>>>>>>>> ] 86/256, 0.3 task/s, elapsed: 330s, ETA: 651s
[>>>>>>>>>> ] 87/256, 0.3 task/s, elapsed: 330s, ETA: 640s
[>>>>>>>>>> ] 88/256, 0.3 task/s, elapsed: 330s, ETA: 630s
[>>>>>>>>>> ] 89/256, 0.3 task/s, elapsed: 331s, ETA: 622s
[>>>>>>>>>> ] 90/256, 0.3 task/s, elapsed: 332s, ETA: 612s
[>>>>>>>>>> ] 91/256, 0.3 task/s, elapsed: 333s, ETA: 604s
[>>>>>>>>>> ] 92/256, 0.3 task/s, elapsed: 334s, ETA: 595s
[>>>>>>>>>> ] 93/256, 0.3 task/s, elapsed: 335s, ETA: 586s
[>>>>>>>>>>> ] 94/256, 0.3 task/s, elapsed: 335s, ETA: 578s
[>>>>>>>>>>> ] 95/256, 0.3 task/s, elapsed: 338s, ETA: 573s
[>>>>>>>>>>> ] 96/256, 0.3 task/s, elapsed: 339s, ETA: 565s
[>>>>>>>>>>> ] 97/256, 0.3 task/s, elapsed: 340s, ETA: 557s
[>>>>>>>>>>> ] 98/256, 0.3 task/s, elapsed: 340s, ETA: 548s
[>>>>>>>>>>> ] 99/256, 0.3 task/s, elapsed: 341s, ETA: 540s
[>>>>>>>>>>> ] 100/256, 0.3 task/s, elapsed: 343s, ETA: 535s
[>>>>>>>>>>> ] 101/256, 0.3 task/s, elapsed: 343s, ETA: 527s
[>>>>>>>>>>> ] 102/256, 0.3 task/s, elapsed: 344s, ETA: 519s
[>>>>>>>>>>> ] 103/256, 0.3 task/s, elapsed: 345s, ETA: 512s
[>>>>>>>>>>> ] 104/256, 0.3 task/s, elapsed: 345s, ETA: 505s
[>>>>>>>>>>> ] 105/256, 0.3 task/s, elapsed: 346s, ETA: 498s
[>>>>>>>>>>>> ] 106/256, 0.3 task/s, elapsed: 347s, ETA: 491s
[>>>>>>>>>>>> ] 107/256, 0.3 task/s, elapsed: 348s, ETA: 485s
[>>>>>>>>>>>> ] 108/256, 0.3 task/s, elapsed: 349s, ETA: 478s
[>>>>>>>>>>>> ] 109/256, 0.3 task/s, elapsed: 350s, ETA: 471s
[>>>>>>>>>>>> ] 110/256, 0.3 task/s, elapsed: 350s, ETA: 464s
[>>>>>>>>>>>> ] 111/256, 0.3 task/s, elapsed: 350s, ETA: 457s
[>>>>>>>>>>>> ] 112/256, 0.3 task/s, elapsed: 351s, ETA: 452s
[>>>>>>>>>>>> ] 113/256, 0.3 task/s, elapsed: 352s, ETA: 445s
[>>>>>>>>>>>> ] 114/256, 0.3 task/s, elapsed: 354s, ETA: 441s
[>>>>>>>>>>>>> ] 115/256, 0.3 task/s, elapsed: 354s, ETA: 434s
[>>>>>>>>>>>>> ] 116/256, 0.3 task/s, elapsed: 354s, ETA: 427s
[>>>>>>>>>>>>> ] 117/256, 0.3 task/s, elapsed: 355s, ETA: 422s
[>>>>>>>>>>>>> ] 118/256, 0.3 task/s, elapsed: 355s, ETA: 416s
[>>>>>>>>>>>>> ] 119/256, 0.3 task/s, elapsed: 356s, ETA: 410s
[>>>>>>>>>>>>> ] 120/256, 0.3 task/s, elapsed: 357s, ETA: 405s
[>>>>>>>>>>>>> ] 121/256, 0.3 task/s, elapsed: 358s, ETA: 399s
[>>>>>>>>>>>>> ] 122/256, 0.3 task/s, elapsed: 358s, ETA: 393s
[>>>>>>>>>>>>> ] 123/256, 0.3 task/s, elapsed: 359s, ETA: 388s
[>>>>>>>>>>>>>> ] 124/256, 0.3 task/s, elapsed: 360s, ETA: 384s
[>>>>>>>>>>>>>> ] 125/256, 0.3 task/s, elapsed: 361s, ETA: 378s
[>>>>>>>>>>>>>> ] 126/256, 0.3 task/s, elapsed: 361s, ETA: 372s
[>>>>>>>>>>>>>> ] 127/256, 0.4 task/s, elapsed: 362s, ETA: 367s
[>>>>>>>>>>>>>> ] 128/256, 0.4 task/s, elapsed: 362s, ETA: 362s
[>>>>>>>>>>>>>> ] 129/256, 0.4 task/s, elapsed: 365s, ETA: 359s
[>>>>>>>>>>>>>> ] 130/256, 0.4 task/s, elapsed: 366s, ETA: 355s
[>>>>>>>>>>>>>> ] 131/256, 0.4 task/s, elapsed: 367s, ETA: 350s
[>>>>>>>>>>>>>> ] 132/256, 0.4 task/s, elapsed: 367s, ETA: 345s
[>>>>>>>>>>>>>>> ] 133/256, 0.4 task/s, elapsed: 367s, ETA: 339s
[>>>>>>>>>>>>>>> ] 134/256, 0.4 task/s, elapsed: 367s, ETA: 334s
[>>>>>>>>>>>>>>> ] 135/256, 0.4 task/s, elapsed: 368s, ETA: 330s
[>>>>>>>>>>>>>>> ] 136/256, 0.4 task/s, elapsed: 369s, ETA: 326s
[>>>>>>>>>>>>>>> ] 137/256, 0.4 task/s, elapsed: 369s, ETA: 321s
[>>>>>>>>>>>>>>> ] 138/256, 0.4 task/s, elapsed: 369s, ETA: 316s
[>>>>>>>>>>>>>>> ] 139/256, 0.4 task/s, elapsed: 370s, ETA: 311s
[>>>>>>>>>>>>>>> ] 140/256, 0.4 task/s, elapsed: 371s, ETA: 307s
[>>>>>>>>>>>>>>> ] 141/256, 0.4 task/s, elapsed: 372s, ETA: 303s
[>>>>>>>>>>>>>>>> ] 142/256, 0.4 task/s, elapsed: 373s, ETA: 300s
[>>>>>>>>>>>>>>>> ] 143/256, 0.4 task/s, elapsed: 373s, ETA: 295s
[>>>>>>>>>>>>>>>> ] 144/256, 0.4 task/s, elapsed: 375s, ETA: 292s
[>>>>>>>>>>>>>>>> ] 145/256, 0.4 task/s, elapsed: 375s, ETA: 287s
[>>>>>>>>>>>>>>>> ] 146/256, 0.4 task/s, elapsed: 376s, ETA: 284s
[>>>>>>>>>>>>>>>> ] 147/256, 0.4 task/s, elapsed: 377s, ETA: 279s
[>>>>>>>>>>>>>>>> ] 148/256, 0.4 task/s, elapsed: 377s, ETA: 275s
[>>>>>>>>>>>>>>>> ] 149/256, 0.4 task/s, elapsed: 377s, ETA: 270s
[>>>>>>>>>>>>>>>> ] 150/256, 0.4 task/s, elapsed: 377s, ETA: 266s
[>>>>>>>>>>>>>>>>> ] 151/256, 0.4 task/s, elapsed: 378s, ETA: 263s
[>>>>>>>>>>>>>>>>> ] 152/256, 0.4 task/s, elapsed: 378s, ETA: 259s
[>>>>>>>>>>>>>>>>> ] 153/256, 0.4 task/s, elapsed: 378s, ETA: 255s
[>>>>>>>>>>>>>>>>> ] 154/256, 0.4 task/s, elapsed: 378s, ETA: 251s
[>>>>>>>>>>>>>>>>> ] 155/256, 0.4 task/s, elapsed: 379s, ETA: 247s
[>>>>>>>>>>>>>>>>> ] 156/256, 0.4 task/s, elapsed: 379s, ETA: 243s
[>>>>>>>>>>>>>>>>> ] 157/256, 0.4 task/s, elapsed: 380s, ETA: 239s
[>>>>>>>>>>>>>>>>> ] 158/256, 0.4 task/s, elapsed: 380s, ETA: 236s
[>>>>>>>>>>>>>>>>>> ] 159/256, 0.4 task/s, elapsed: 380s, ETA: 232s
[>>>>>>>>>>>>>>>>>> ] 160/256, 0.4 task/s, elapsed: 380s, ETA: 228s
[>>>>>>>>>>>>>>>>>> ] 161/256, 0.4 task/s, elapsed: 380s, ETA: 224s
[>>>>>>>>>>>>>>>>>> ] 162/256, 0.4 task/s, elapsed: 380s, ETA: 221s
[>>>>>>>>>>>>>>>>>> ] 163/256, 0.4 task/s, elapsed: 380s, ETA: 217s
[>>>>>>>>>>>>>>>>>> ] 164/256, 0.4 task/s, elapsed: 381s, ETA: 214s
[>>>>>>>>>>>>>>>>>> ] 165/256, 0.4 task/s, elapsed: 384s, ETA: 212s
[>>>>>>>>>>>>>>>>>> ] 166/256, 0.4 task/s, elapsed: 385s, ETA: 209s
[>>>>>>>>>>>>>>>>>> ] 167/256, 0.4 task/s, elapsed: 385s, ETA: 205s
[>>>>>>>>>>>>>>>>>>> ] 168/256, 0.4 task/s, elapsed: 385s, ETA: 202s
[>>>>>>>>>>>>>>>>>>> ] 169/256, 0.4 task/s, elapsed: 386s, ETA: 199s
[>>>>>>>>>>>>>>>>>>> ] 170/256, 0.4 task/s, elapsed: 386s, ETA: 195s
[>>>>>>>>>>>>>>>>>>> ] 171/256, 0.4 task/s, elapsed: 388s, ETA: 193s
[>>>>>>>>>>>>>>>>>>> ] 172/256, 0.4 task/s, elapsed: 388s, ETA: 190s
[>>>>>>>>>>>>>>>>>>> ] 173/256, 0.4 task/s, elapsed: 389s, ETA: 187s
[>>>>>>>>>>>>>>>>>>> ] 174/256, 0.4 task/s, elapsed: 390s, ETA: 184s
[>>>>>>>>>>>>>>>>>>> ] 175/256, 0.4 task/s, elapsed: 390s, ETA: 181s
[>>>>>>>>>>>>>>>>>>> ] 176/256, 0.5 task/s, elapsed: 391s, ETA: 178s
[>>>>>>>>>>>>>>>>>>>> ] 177/256, 0.5 task/s, elapsed: 391s, ETA: 175s
[>>>>>>>>>>>>>>>>>>>> ] 178/256, 0.5 task/s, elapsed: 391s, ETA: 172s
[>>>>>>>>>>>>>>>>>>>> ] 179/256, 0.5 task/s, elapsed: 392s, ETA: 169s
[>>>>>>>>>>>>>>>>>>>> ] 180/256, 0.5 task/s, elapsed: 393s, ETA: 166s
[>>>>>>>>>>>>>>>>>>>> ] 181/256, 0.5 task/s, elapsed: 394s, ETA: 163s
[>>>>>>>>>>>>>>>>>>>> ] 182/256, 0.5 task/s, elapsed: 395s, ETA: 160s
[>>>>>>>>>>>>>>>>>>>> ] 183/256, 0.5 task/s, elapsed: 396s, ETA: 158s
[>>>>>>>>>>>>>>>>>>>> ] 184/256, 0.5 task/s, elapsed: 396s, ETA: 155s
[>>>>>>>>>>>>>>>>>>>> ] 185/256, 0.5 task/s, elapsed: 397s, ETA: 152s
[>>>>>>>>>>>>>>>>>>>>> ] 186/256, 0.5 task/s, elapsed: 397s, ETA: 149s
[>>>>>>>>>>>>>>>>>>>>> ] 187/256, 0.5 task/s, elapsed: 397s, ETA: 147s
[>>>>>>>>>>>>>>>>>>>>> ] 188/256, 0.5 task/s, elapsed: 398s, ETA: 144s
[>>>>>>>>>>>>>>>>>>>>> ] 189/256, 0.5 task/s, elapsed: 399s, ETA: 141s
[>>>>>>>>>>>>>>>>>>>>> ] 190/256, 0.5 task/s, elapsed: 400s, ETA: 139s
[>>>>>>>>>>>>>>>>>>>>> ] 191/256, 0.5 task/s, elapsed: 400s, ETA: 136s
[>>>>>>>>>>>>>>>>>>>>> ] 192/256, 0.5 task/s, elapsed: 401s, ETA: 134s
[>>>>>>>>>>>>>>>>>>>>> ] 193/256, 0.5 task/s, elapsed: 401s, ETA: 131s
[>>>>>>>>>>>>>>>>>>>>> ] 194/256, 0.5 task/s, elapsed: 402s, ETA: 128s
[>>>>>>>>>>>>>>>>>>>>>> ] 195/256, 0.5 task/s, elapsed: 402s, ETA: 126s
[>>>>>>>>>>>>>>>>>>>>>> ] 196/256, 0.5 task/s, elapsed: 403s, ETA: 123s
[>>>>>>>>>>>>>>>>>>>>>> ] 197/256, 0.5 task/s, elapsed: 403s, ETA: 121s
[>>>>>>>>>>>>>>>>>>>>>> ] 198/256, 0.5 task/s, elapsed: 403s, ETA: 118s
[>>>>>>>>>>>>>>>>>>>>>> ] 199/256, 0.5 task/s, elapsed: 404s, ETA: 116s
[>>>>>>>>>>>>>>>>>>>>>> ] 200/256, 0.5 task/s, elapsed: 405s, ETA: 113s
[>>>>>>>>>>>>>>>>>>>>>> ] 201/256, 0.5 task/s, elapsed: 407s, ETA: 111s
[>>>>>>>>>>>>>>>>>>>>>> ] 202/256, 0.5 task/s, elapsed: 407s, ETA: 109s
[>>>>>>>>>>>>>>>>>>>>>> ] 203/256, 0.5 task/s, elapsed: 408s, ETA: 107s
[>>>>>>>>>>>>>>>>>>>>>>> ] 204/256, 0.5 task/s, elapsed: 408s, ETA: 104s
[>>>>>>>>>>>>>>>>>>>>>>> ] 205/256, 0.5 task/s, elapsed: 408s, ETA: 102s
[>>>>>>>>>>>>>>>>>>>>>>> ] 206/256, 0.5 task/s, elapsed: 410s, ETA: 99s
[>>>>>>>>>>>>>>>>>>>>>>> ] 207/256, 0.5 task/s, elapsed: 410s, ETA: 97s
[>>>>>>>>>>>>>>>>>>>>>>> ] 208/256, 0.5 task/s, elapsed: 410s, ETA: 95s
[>>>>>>>>>>>>>>>>>>>>>>> ] 209/256, 0.5 task/s, elapsed: 410s, ETA: 92s
[>>>>>>>>>>>>>>>>>>>>>>> ] 210/256, 0.5 task/s, elapsed: 411s, ETA: 90s
[>>>>>>>>>>>>>>>>>>>>>>> ] 211/256, 0.5 task/s, elapsed: 412s, ETA: 88s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 212/256, 0.5 task/s, elapsed: 413s, ETA: 86s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 213/256, 0.5 task/s, elapsed: 413s, ETA: 83s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 214/256, 0.5 task/s, elapsed: 415s, ETA: 81s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 215/256, 0.5 task/s, elapsed: 415s, ETA: 79s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 216/256, 0.5 task/s, elapsed: 415s, ETA: 77s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 217/256, 0.5 task/s, elapsed: 417s, ETA: 75s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 218/256, 0.5 task/s, elapsed: 417s, ETA: 73s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 219/256, 0.5 task/s, elapsed: 417s, ETA: 71s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 220/256, 0.5 task/s, elapsed: 419s, ETA: 69s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 221/256, 0.5 task/s, elapsed: 420s, ETA: 66s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 222/256, 0.5 task/s, elapsed: 420s, ETA: 64s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 223/256, 0.5 task/s, elapsed: 422s, ETA: 62s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 224/256, 0.5 task/s, elapsed: 423s, ETA: 60s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 225/256, 0.5 task/s, elapsed: 425s, ETA: 58s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 226/256, 0.5 task/s, elapsed: 425s, ETA: 56s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 227/256, 0.5 task/s, elapsed: 428s, ETA: 55s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 228/256, 0.5 task/s, elapsed: 428s, ETA: 53s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 229/256, 0.5 task/s, elapsed: 428s, ETA: 51s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 230/256, 0.5 task/s, elapsed: 429s, ETA: 48s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 231/256, 0.5 task/s, elapsed: 431s, ETA: 47s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 232/256, 0.5 task/s, elapsed: 431s, ETA: 45s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 233/256, 0.5 task/s, elapsed: 431s, ETA: 43s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 234/256, 0.5 task/s, elapsed: 431s, ETA: 41s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 235/256, 0.5 task/s, elapsed: 431s, ETA: 39s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 236/256, 0.5 task/s, elapsed: 433s, ETA: 37s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 237/256, 0.5 task/s, elapsed: 437s, ETA: 35s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 238/256, 0.5 task/s, elapsed: 438s, ETA: 33s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 239/256, 0.5 task/s, elapsed: 442s, ETA: 31s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 240/256, 0.5 task/s, elapsed: 443s, ETA: 30s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 241/256, 0.5 task/s, elapsed: 444s, ETA: 28s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 242/256, 0.5 task/s, elapsed: 444s, ETA: 26s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 243/256, 0.5 task/s, elapsed: 446s, ETA: 24s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 244/256, 0.5 task/s, elapsed: 448s, ETA: 22s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 245/256, 0.5 task/s, elapsed: 452s, ETA: 20s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 246/256, 0.5 task/s, elapsed: 454s, ETA: 18s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 247/256, 0.5 task/s, elapsed: 455s, ETA: 17s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 248/256, 0.5 task/s, elapsed: 457s, ETA: 15s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 249/256, 0.5 task/s, elapsed: 463s, ETA: 13s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 250/256, 0.5 task/s, elapsed: 467s, ETA: 11s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 251/256, 0.5 task/s, elapsed: 468s, ETA: 9s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 252/256, 0.5 task/s, elapsed: 483s, ETA: 8s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 253/256, 0.5 task/s, elapsed: 484s, ETA: 6s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 254/256, 0.5 task/s, elapsed: 487s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 255/256, 0.5 task/s, elapsed: 504s, ETA: 2s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>>] 256/256, 0.5 task/s, elapsed: 519s, ETA: 0s
12/05 03:48:35 - OpenCompass - INFO - Partitioned into 287 tasks.
[ ] 0/287, elapsed: 0s, ETA:
[ ] 1/287, 0.1 task/s, elapsed: 18s, ETA: 5192s
[ ] 2/287, 0.1 task/s, elapsed: 21s, ETA: 3005s
[ ] 3/287, 0.1 task/s, elapsed: 21s, ETA: 2001s
[ ] 4/287, 0.2 task/s, elapsed: 21s, ETA: 1496s
[ ] 5/287, 0.2 task/s, elapsed: 21s, ETA: 1201s
[ ] 6/287, 0.3 task/s, elapsed: 22s, ETA: 1009s
[ ] 7/287, 0.3 task/s, elapsed: 25s, ETA: 1001s
[ ] 8/287, 0.3 task/s, elapsed: 25s, ETA: 879s
[> ] 9/287, 0.4 task/s, elapsed: 25s, ETA: 780s
[> ] 10/287, 0.4 task/s, elapsed: 25s, ETA: 700s
[> ] 11/287, 0.4 task/s, elapsed: 25s, ETA: 634s
[> ] 12/287, 0.5 task/s, elapsed: 25s, ETA: 580s
[> ] 13/287, 0.5 task/s, elapsed: 25s, ETA: 534s
[> ] 14/287, 0.6 task/s, elapsed: 25s, ETA: 494s
[> ] 15/287, 0.6 task/s, elapsed: 25s, ETA: 460s
[> ] 16/287, 0.6 task/s, elapsed: 25s, ETA: 429s
[> ] 17/287, 0.7 task/s, elapsed: 25s, ETA: 403s
[> ] 18/287, 0.7 task/s, elapsed: 25s, ETA: 379s
[>> ] 19/287, 0.7 task/s, elapsed: 25s, ETA: 358s
[>> ] 20/287, 0.8 task/s, elapsed: 25s, ETA: 339s
[>> ] 21/287, 0.8 task/s, elapsed: 25s, ETA: 322s
[>> ] 22/287, 0.9 task/s, elapsed: 25s, ETA: 306s
[>> ] 23/287, 0.9 task/s, elapsed: 26s, ETA: 293s
[>> ] 24/287, 0.9 task/s, elapsed: 26s, ETA: 280s
[>> ] 25/287, 0.8 task/s, elapsed: 30s, ETA: 319s
[>> ] 26/287, 0.9 task/s, elapsed: 30s, ETA: 306s
[>> ] 27/287, 0.9 task/s, elapsed: 31s, ETA: 294s
[>>> ] 28/287, 0.9 task/s, elapsed: 31s, ETA: 283s
[>>> ] 29/287, 0.9 task/s, elapsed: 31s, ETA: 272s
[>>> ] 30/287, 1.0 task/s, elapsed: 31s, ETA: 262s
[>>> ] 31/287, 1.0 task/s, elapsed: 31s, ETA: 253s
[>>> ] 32/287, 1.0 task/s, elapsed: 31s, ETA: 244s
[>>> ] 33/287, 1.1 task/s, elapsed: 31s, ETA: 236s
[>>> ] 34/287, 1.1 task/s, elapsed: 31s, ETA: 228s
[>>> ] 35/287, 1.1 task/s, elapsed: 31s, ETA: 221s
[>>> ] 36/287, 1.2 task/s, elapsed: 31s, ETA: 214s
[>>> ] 37/287, 1.2 task/s, elapsed: 31s, ETA: 207s
[>>>> ] 38/287, 1.2 task/s, elapsed: 31s, ETA: 201s
[>>>> ] 39/287, 1.3 task/s, elapsed: 31s, ETA: 195s
[>>>> ] 40/287, 1.3 task/s, elapsed: 31s, ETA: 190s
[>>>> ] 41/287, 1.3 task/s, elapsed: 31s, ETA: 184s
[>>>> ] 42/287, 1.4 task/s, elapsed: 31s, ETA: 179s
[>>>> ] 43/287, 1.4 task/s, elapsed: 31s, ETA: 174s
[>>>> ] 44/287, 1.4 task/s, elapsed: 31s, ETA: 170s
[>>>> ] 45/287, 1.5 task/s, elapsed: 31s, ETA: 166s
[>>>> ] 46/287, 1.5 task/s, elapsed: 31s, ETA: 161s
[>>>>> ] 47/287, 1.5 task/s, elapsed: 31s, ETA: 158s
[>>>>> ] 48/287, 1.6 task/s, elapsed: 31s, ETA: 154s
[>>>>> ] 49/287, 1.6 task/s, elapsed: 31s, ETA: 152s
[>>>>> ] 50/287, 1.6 task/s, elapsed: 31s, ETA: 149s
[>>>>> ] 51/287, 1.6 task/s, elapsed: 32s, ETA: 146s
[>>>>> ] 52/287, 1.6 task/s, elapsed: 32s, ETA: 143s
[>>>>> ] 53/287, 1.7 task/s, elapsed: 32s, ETA: 140s
[>>>>> ] 54/287, 1.7 task/s, elapsed: 32s, ETA: 137s
[>>>>> ] 55/287, 1.7 task/s, elapsed: 32s, ETA: 134s
[>>>>>> ] 56/287, 1.8 task/s, elapsed: 32s, ETA: 131s
[>>>>>> ] 57/287, 1.8 task/s, elapsed: 32s, ETA: 129s
[>>>>>> ] 58/287, 1.8 task/s, elapsed: 32s, ETA: 126s
[>>>>>> ] 59/287, 1.8 task/s, elapsed: 32s, ETA: 124s
[>>>>>> ] 60/287, 1.9 task/s, elapsed: 32s, ETA: 121s
[>>>>>> ] 61/287, 1.9 task/s, elapsed: 32s, ETA: 119s
[>>>>>> ] 62/287, 1.9 task/s, elapsed: 32s, ETA: 117s
[>>>>>> ] 63/287, 2.0 task/s, elapsed: 32s, ETA: 114s
[>>>>>> ] 64/287, 2.0 task/s, elapsed: 33s, ETA: 114s
[>>>>>>> ] 65/287, 2.0 task/s, elapsed: 33s, ETA: 112s
[>>>>>>> ] 66/287, 2.0 task/s, elapsed: 33s, ETA: 110s
[>>>>>>> ] 67/287, 2.0 task/s, elapsed: 33s, ETA: 108s
[>>>>>>> ] 68/287, 2.1 task/s, elapsed: 33s, ETA: 106s
[>>>>>>> ] 69/287, 2.1 task/s, elapsed: 33s, ETA: 104s
[>>>>>>> ] 70/287, 2.1 task/s, elapsed: 33s, ETA: 103s
[>>>>>>> ] 71/287, 2.1 task/s, elapsed: 33s, ETA: 101s
[>>>>>>> ] 72/287, 2.1 task/s, elapsed: 34s, ETA: 100s
[>>>>>>> ] 73/287, 2.1 task/s, elapsed: 35s, ETA: 102s
[>>>>>>> ] 74/287, 2.1 task/s, elapsed: 35s, ETA: 100s
[>>>>>>>> ] 75/287, 2.1 task/s, elapsed: 35s, ETA: 99s
[>>>>>>>> ] 76/287, 2.2 task/s, elapsed: 35s, ETA: 97s
[>>>>>>>> ] 77/287, 2.2 task/s, elapsed: 35s, ETA: 96s
[>>>>>>>> ] 78/287, 2.2 task/s, elapsed: 35s, ETA: 94s
[>>>>>>>> ] 79/287, 2.2 task/s, elapsed: 35s, ETA: 92s
[>>>>>>>> ] 80/287, 2.3 task/s, elapsed: 35s, ETA: 91s
[>>>>>>>> ] 81/287, 2.3 task/s, elapsed: 35s, ETA: 90s
[>>>>>>>> ] 82/287, 2.3 task/s, elapsed: 35s, ETA: 88s
[>>>>>>>> ] 83/287, 2.3 task/s, elapsed: 35s, ETA: 87s
[>>>>>>>>> ] 84/287, 2.4 task/s, elapsed: 35s, ETA: 86s
[>>>>>>>>> ] 85/287, 2.4 task/s, elapsed: 35s, ETA: 84s
[>>>>>>>>> ] 86/287, 2.4 task/s, elapsed: 35s, ETA: 83s
[>>>>>>>>> ] 87/287, 2.5 task/s, elapsed: 35s, ETA: 82s
[>>>>>>>>> ] 88/287, 2.5 task/s, elapsed: 35s, ETA: 80s
[>>>>>>>>> ] 89/287, 2.5 task/s, elapsed: 35s, ETA: 79s
[>>>>>>>>> ] 90/287, 2.5 task/s, elapsed: 36s, ETA: 78s
[>>>>>>>>> ] 91/287, 2.6 task/s, elapsed: 36s, ETA: 77s
[>>>>>>>>> ] 92/287, 2.6 task/s, elapsed: 36s, ETA: 75s
[>>>>>>>>>> ] 93/287, 2.6 task/s, elapsed: 36s, ETA: 74s
[>>>>>>>>>> ] 94/287, 2.6 task/s, elapsed: 36s, ETA: 73s
[>>>>>>>>>> ] 95/287, 2.7 task/s, elapsed: 36s, ETA: 72s
[>>>>>>>>>> ] 96/287, 2.7 task/s, elapsed: 36s, ETA: 71s
[>>>>>>>>>> ] 97/287, 2.7 task/s, elapsed: 36s, ETA: 70s
[>>>>>>>>>> ] 98/287, 2.7 task/s, elapsed: 36s, ETA: 70s
[>>>>>>>>>> ] 99/287, 2.7 task/s, elapsed: 36s, ETA: 69s
[>>>>>>>>>> ] 100/287, 2.8 task/s, elapsed: 36s, ETA: 68s
[>>>>>>>>>> ] 101/287, 2.8 task/s, elapsed: 36s, ETA: 67s
[>>>>>>>>>> ] 102/287, 2.8 task/s, elapsed: 36s, ETA: 66s
[>>>>>>>>>> ] 103/287, 2.9 task/s, elapsed: 36s, ETA: 65s
[>>>>>>>>>> ] 104/287, 2.9 task/s, elapsed: 36s, ETA: 64s
[>>>>>>>>>> ] 105/287, 2.9 task/s, elapsed: 36s, ETA: 63s
[>>>>>>>>>>> ] 106/287, 2.9 task/s, elapsed: 36s, ETA: 62s
[>>>>>>>>>>> ] 107/287, 3.0 task/s, elapsed: 36s, ETA: 61s
[>>>>>>>>>>> ] 108/287, 3.0 task/s, elapsed: 36s, ETA: 60s
[>>>>>>>>>>> ] 109/287, 3.0 task/s, elapsed: 36s, ETA: 59s
[>>>>>>>>>>> ] 110/287, 3.0 task/s, elapsed: 36s, ETA: 58s
[>>>>>>>>>>> ] 111/287, 3.1 task/s, elapsed: 36s, ETA: 58s
[>>>>>>>>>>> ] 112/287, 3.1 task/s, elapsed: 36s, ETA: 57s
[>>>>>>>>>>> ] 113/287, 3.1 task/s, elapsed: 36s, ETA: 56s
[>>>>>>>>>>> ] 114/287, 3.1 task/s, elapsed: 36s, ETA: 55s
[>>>>>>>>>>>> ] 115/287, 3.2 task/s, elapsed: 36s, ETA: 54s
[>>>>>>>>>>>> ] 116/287, 3.2 task/s, elapsed: 36s, ETA: 54s
[>>>>>>>>>>>> ] 117/287, 3.2 task/s, elapsed: 36s, ETA: 53s
[>>>>>>>>>>>> ] 118/287, 3.2 task/s, elapsed: 36s, ETA: 52s
[>>>>>>>>>>>> ] 119/287, 3.3 task/s, elapsed: 36s, ETA: 51s
[>>>>>>>>>>>> ] 120/287, 3.3 task/s, elapsed: 36s, ETA: 51s
[>>>>>>>>>>>> ] 121/287, 3.3 task/s, elapsed: 37s, ETA: 50s
[>>>>>>>>>>>> ] 122/287, 3.3 task/s, elapsed: 37s, ETA: 49s
[>>>>>>>>>>>> ] 123/287, 3.4 task/s, elapsed: 37s, ETA: 49s
[>>>>>>>>>>>> ] 124/287, 3.4 task/s, elapsed: 37s, ETA: 48s
[>>>>>>>>>>>>> ] 125/287, 3.4 task/s, elapsed: 37s, ETA: 48s
[>>>>>>>>>>>>> ] 126/287, 3.4 task/s, elapsed: 37s, ETA: 47s
[>>>>>>>>>>>>> ] 127/287, 3.4 task/s, elapsed: 37s, ETA: 47s
[>>>>>>>>>>>>> ] 128/287, 3.4 task/s, elapsed: 37s, ETA: 46s
[>>>>>>>>>>>>> ] 129/287, 3.4 task/s, elapsed: 37s, ETA: 46s
[>>>>>>>>>>>>> ] 130/287, 3.5 task/s, elapsed: 38s, ETA: 45s
[>>>>>>>>>>>>> ] 131/287, 3.5 task/s, elapsed: 38s, ETA: 45s
[>>>>>>>>>>>>> ] 132/287, 3.5 task/s, elapsed: 38s, ETA: 44s
[>>>>>>>>>>>>> ] 133/287, 3.5 task/s, elapsed: 38s, ETA: 44s
[>>>>>>>>>>>>>> ] 134/287, 3.5 task/s, elapsed: 38s, ETA: 43s
[>>>>>>>>>>>>>> ] 135/287, 3.6 task/s, elapsed: 38s, ETA: 43s
[>>>>>>>>>>>>>> ] 136/287, 3.6 task/s, elapsed: 38s, ETA: 42s
[>>>>>>>>>>>>>> ] 137/287, 3.6 task/s, elapsed: 38s, ETA: 42s
[>>>>>>>>>>>>>> ] 138/287, 3.6 task/s, elapsed: 38s, ETA: 41s
[>>>>>>>>>>>>>> ] 139/287, 3.7 task/s, elapsed: 38s, ETA: 40s
[>>>>>>>>>>>>>> ] 140/287, 3.7 task/s, elapsed: 38s, ETA: 40s
[>>>>>>>>>>>>>> ] 141/287, 3.7 task/s, elapsed: 38s, ETA: 39s
[>>>>>>>>>>>>>> ] 142/287, 3.7 task/s, elapsed: 38s, ETA: 39s
[>>>>>>>>>>>>>> ] 143/287, 3.8 task/s, elapsed: 38s, ETA: 38s
[>>>>>>>>>>>>>>> ] 144/287, 3.8 task/s, elapsed: 38s, ETA: 38s
[>>>>>>>>>>>>>>> ] 145/287, 3.8 task/s, elapsed: 38s, ETA: 37s
[>>>>>>>>>>>>>>> ] 146/287, 3.8 task/s, elapsed: 38s, ETA: 37s
[>>>>>>>>>>>>>>> ] 147/287, 3.9 task/s, elapsed: 38s, ETA: 36s
[>>>>>>>>>>>>>>> ] 148/287, 3.9 task/s, elapsed: 38s, ETA: 36s
[>>>>>>>>>>>>>>> ] 149/287, 3.9 task/s, elapsed: 38s, ETA: 35s
[>>>>>>>>>>>>>>> ] 150/287, 3.9 task/s, elapsed: 38s, ETA: 35s
[>>>>>>>>>>>>>>> ] 151/287, 4.0 task/s, elapsed: 38s, ETA: 34s
[>>>>>>>>>>>>>>> ] 152/287, 4.0 task/s, elapsed: 38s, ETA: 34s
[>>>>>>>>>>>>>>> ] 153/287, 4.0 task/s, elapsed: 38s, ETA: 33s
[>>>>>>>>>>>>>>>> ] 154/287, 4.0 task/s, elapsed: 38s, ETA: 33s
[>>>>>>>>>>>>>>>> ] 155/287, 4.1 task/s, elapsed: 38s, ETA: 33s
[>>>>>>>>>>>>>>>> ] 156/287, 4.1 task/s, elapsed: 38s, ETA: 32s
[>>>>>>>>>>>>>>>> ] 157/287, 4.1 task/s, elapsed: 38s, ETA: 32s
[>>>>>>>>>>>>>>>> ] 158/287, 4.1 task/s, elapsed: 38s, ETA: 31s
[>>>>>>>>>>>>>>>> ] 159/287, 4.2 task/s, elapsed: 38s, ETA: 31s
[>>>>>>>>>>>>>>>> ] 160/287, 4.2 task/s, elapsed: 38s, ETA: 30s
[>>>>>>>>>>>>>>>> ] 161/287, 4.2 task/s, elapsed: 38s, ETA: 30s
[>>>>>>>>>>>>>>>> ] 162/287, 4.2 task/s, elapsed: 38s, ETA: 30s
[>>>>>>>>>>>>>>>>> ] 163/287, 4.2 task/s, elapsed: 38s, ETA: 29s
[>>>>>>>>>>>>>>>>> ] 164/287, 4.3 task/s, elapsed: 38s, ETA: 29s
[>>>>>>>>>>>>>>>>> ] 165/287, 4.3 task/s, elapsed: 38s, ETA: 28s
[>>>>>>>>>>>>>>>>> ] 166/287, 4.3 task/s, elapsed: 38s, ETA: 28s
[>>>>>>>>>>>>>>>>> ] 167/287, 4.3 task/s, elapsed: 38s, ETA: 28s
[>>>>>>>>>>>>>>>>> ] 168/287, 4.4 task/s, elapsed: 38s, ETA: 27s
[>>>>>>>>>>>>>>>>> ] 169/287, 4.4 task/s, elapsed: 38s, ETA: 27s
[>>>>>>>>>>>>>>>>> ] 170/287, 4.4 task/s, elapsed: 38s, ETA: 26s
[>>>>>>>>>>>>>>>>> ] 171/287, 4.4 task/s, elapsed: 39s, ETA: 26s
[>>>>>>>>>>>>>>>>> ] 172/287, 4.4 task/s, elapsed: 39s, ETA: 26s
[>>>>>>>>>>>>>>>>>> ] 173/287, 4.4 task/s, elapsed: 39s, ETA: 26s
[>>>>>>>>>>>>>>>>>> ] 174/287, 4.4 task/s, elapsed: 39s, ETA: 25s
[>>>>>>>>>>>>>>>>>> ] 175/287, 4.4 task/s, elapsed: 39s, ETA: 25s
[>>>>>>>>>>>>>>>>>> ] 176/287, 4.5 task/s, elapsed: 39s, ETA: 25s
[>>>>>>>>>>>>>>>>>> ] 177/287, 4.5 task/s, elapsed: 39s, ETA: 24s
[>>>>>>>>>>>>>>>>>> ] 178/287, 4.5 task/s, elapsed: 39s, ETA: 24s
[>>>>>>>>>>>>>>>>>> ] 179/287, 4.5 task/s, elapsed: 39s, ETA: 24s
[>>>>>>>>>>>>>>>>>> ] 180/287, 4.6 task/s, elapsed: 39s, ETA: 23s
[>>>>>>>>>>>>>>>>>> ] 181/287, 4.6 task/s, elapsed: 39s, ETA: 23s
[>>>>>>>>>>>>>>>>>>> ] 182/287, 4.6 task/s, elapsed: 39s, ETA: 23s
[>>>>>>>>>>>>>>>>>>> ] 183/287, 4.6 task/s, elapsed: 39s, ETA: 22s
[>>>>>>>>>>>>>>>>>>> ] 184/287, 4.7 task/s, elapsed: 39s, ETA: 22s
[>>>>>>>>>>>>>>>>>>> ] 185/287, 4.7 task/s, elapsed: 39s, ETA: 22s
[>>>>>>>>>>>>>>>>>>> ] 186/287, 4.7 task/s, elapsed: 39s, ETA: 21s
[>>>>>>>>>>>>>>>>>>> ] 187/287, 4.7 task/s, elapsed: 39s, ETA: 21s
[>>>>>>>>>>>>>>>>>>> ] 188/287, 4.8 task/s, elapsed: 40s, ETA: 21s
[>>>>>>>>>>>>>>>>>>> ] 189/287, 4.8 task/s, elapsed: 40s, ETA: 20s
[>>>>>>>>>>>>>>>>>>> ] 190/287, 4.8 task/s, elapsed: 40s, ETA: 20s
[>>>>>>>>>>>>>>>>>>> ] 191/287, 4.8 task/s, elapsed: 40s, ETA: 20s
[>>>>>>>>>>>>>>>>>>>> ] 192/287, 4.8 task/s, elapsed: 40s, ETA: 20s
[>>>>>>>>>>>>>>>>>>>> ] 193/287, 4.9 task/s, elapsed: 40s, ETA: 19s
[>>>>>>>>>>>>>>>>>>>> ] 194/287, 4.8 task/s, elapsed: 40s, ETA: 19s
[>>>>>>>>>>>>>>>>>>>> ] 195/287, 4.8 task/s, elapsed: 40s, ETA: 19s
[>>>>>>>>>>>>>>>>>>>> ] 196/287, 4.8 task/s, elapsed: 40s, ETA: 19s
[>>>>>>>>>>>>>>>>>>>> ] 197/287, 4.9 task/s, elapsed: 40s, ETA: 18s
[>>>>>>>>>>>>>>>>>>>> ] 198/287, 4.9 task/s, elapsed: 41s, ETA: 18s
[>>>>>>>>>>>>>>>>>>>> ] 199/287, 4.9 task/s, elapsed: 41s, ETA: 18s
[>>>>>>>>>>>>>>>>>>>> ] 200/287, 4.9 task/s, elapsed: 41s, ETA: 18s
[>>>>>>>>>>>>>>>>>>>>> ] 201/287, 5.0 task/s, elapsed: 41s, ETA: 17s
[>>>>>>>>>>>>>>>>>>>>> ] 202/287, 5.0 task/s, elapsed: 41s, ETA: 17s
[>>>>>>>>>>>>>>>>>>>>> ] 203/287, 5.0 task/s, elapsed: 41s, ETA: 17s
[>>>>>>>>>>>>>>>>>>>>> ] 204/287, 5.0 task/s, elapsed: 41s, ETA: 17s
[>>>>>>>>>>>>>>>>>>>>> ] 205/287, 5.0 task/s, elapsed: 41s, ETA: 16s
[>>>>>>>>>>>>>>>>>>>>> ] 206/287, 5.1 task/s, elapsed: 41s, ETA: 16s
[>>>>>>>>>>>>>>>>>>>>> ] 207/287, 5.0 task/s, elapsed: 41s, ETA: 16s
[>>>>>>>>>>>>>>>>>>>>> ] 208/287, 5.1 task/s, elapsed: 41s, ETA: 16s
[>>>>>>>>>>>>>>>>>>>>> ] 209/287, 5.1 task/s, elapsed: 41s, ETA: 15s
[>>>>>>>>>>>>>>>>>>>>> ] 210/287, 5.1 task/s, elapsed: 41s, ETA: 15s
[>>>>>>>>>>>>>>>>>>>>>> ] 211/287, 5.1 task/s, elapsed: 42s, ETA: 15s
[>>>>>>>>>>>>>>>>>>>>>> ] 212/287, 5.1 task/s, elapsed: 42s, ETA: 15s
[>>>>>>>>>>>>>>>>>>>>>> ] 213/287, 5.1 task/s, elapsed: 42s, ETA: 15s
[>>>>>>>>>>>>>>>>>>>>>> ] 214/287, 5.1 task/s, elapsed: 42s, ETA: 14s
[>>>>>>>>>>>>>>>>>>>>>> ] 215/287, 5.1 task/s, elapsed: 42s, ETA: 14s
[>>>>>>>>>>>>>>>>>>>>>> ] 216/287, 5.2 task/s, elapsed: 42s, ETA: 14s
[>>>>>>>>>>>>>>>>>>>>>> ] 217/287, 5.1 task/s, elapsed: 42s, ETA: 14s
[>>>>>>>>>>>>>>>>>>>>>> ] 218/287, 5.2 task/s, elapsed: 42s, ETA: 13s
[>>>>>>>>>>>>>>>>>>>>>> ] 219/287, 5.2 task/s, elapsed: 42s, ETA: 13s
[>>>>>>>>>>>>>>>>>>>>>> ] 220/287, 5.2 task/s, elapsed: 42s, ETA: 13s
[>>>>>>>>>>>>>>>>>>>>>>> ] 221/287, 5.2 task/s, elapsed: 42s, ETA: 13s
[>>>>>>>>>>>>>>>>>>>>>>> ] 222/287, 5.2 task/s, elapsed: 42s, ETA: 12s
[>>>>>>>>>>>>>>>>>>>>>>> ] 223/287, 5.2 task/s, elapsed: 42s, ETA: 12s
[>>>>>>>>>>>>>>>>>>>>>>> ] 224/287, 5.3 task/s, elapsed: 42s, ETA: 12s
[>>>>>>>>>>>>>>>>>>>>>>> ] 225/287, 5.3 task/s, elapsed: 43s, ETA: 12s
[>>>>>>>>>>>>>>>>>>>>>>> ] 226/287, 5.3 task/s, elapsed: 43s, ETA: 11s
[>>>>>>>>>>>>>>>>>>>>>>> ] 227/287, 5.3 task/s, elapsed: 43s, ETA: 11s
[>>>>>>>>>>>>>>>>>>>>>>> ] 228/287, 5.3 task/s, elapsed: 43s, ETA: 11s
[>>>>>>>>>>>>>>>>>>>>>>> ] 229/287, 5.4 task/s, elapsed: 43s, ETA: 11s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 230/287, 5.4 task/s, elapsed: 43s, ETA: 11s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 231/287, 5.4 task/s, elapsed: 43s, ETA: 10s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 232/287, 5.4 task/s, elapsed: 43s, ETA: 10s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 233/287, 5.5 task/s, elapsed: 43s, ETA: 10s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 234/287, 5.5 task/s, elapsed: 43s, ETA: 10s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 235/287, 5.5 task/s, elapsed: 43s, ETA: 9s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 236/287, 5.5 task/s, elapsed: 43s, ETA: 9s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 237/287, 5.5 task/s, elapsed: 43s, ETA: 9s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 238/287, 5.6 task/s, elapsed: 43s, ETA: 9s
[>>>>>>>>>>>>>>>>>>>>>>>> ] 239/287, 5.6 task/s, elapsed: 43s, ETA: 9s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 240/287, 5.6 task/s, elapsed: 43s, ETA: 8s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 241/287, 5.6 task/s, elapsed: 43s, ETA: 8s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 242/287, 5.6 task/s, elapsed: 43s, ETA: 8s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 243/287, 5.7 task/s, elapsed: 43s, ETA: 8s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 244/287, 5.7 task/s, elapsed: 43s, ETA: 8s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 245/287, 5.7 task/s, elapsed: 43s, ETA: 7s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 246/287, 5.7 task/s, elapsed: 43s, ETA: 7s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 247/287, 5.7 task/s, elapsed: 43s, ETA: 7s
[>>>>>>>>>>>>>>>>>>>>>>>>> ] 248/287, 5.8 task/s, elapsed: 43s, ETA: 7s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 249/287, 5.8 task/s, elapsed: 43s, ETA: 7s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 250/287, 5.8 task/s, elapsed: 43s, ETA: 6s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 251/287, 5.8 task/s, elapsed: 43s, ETA: 6s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 252/287, 5.9 task/s, elapsed: 43s, ETA: 6s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 253/287, 5.9 task/s, elapsed: 43s, ETA: 6s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 254/287, 5.9 task/s, elapsed: 43s, ETA: 6s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 255/287, 5.9 task/s, elapsed: 43s, ETA: 5s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 256/287, 5.9 task/s, elapsed: 43s, ETA: 5s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 257/287, 6.0 task/s, elapsed: 43s, ETA: 5s
[>>>>>>>>>>>>>>>>>>>>>>>>>> ] 258/287, 6.0 task/s, elapsed: 43s, ETA: 5s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 259/287, 6.0 task/s, elapsed: 43s, ETA: 5s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 260/287, 6.0 task/s, elapsed: 43s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 261/287, 5.9 task/s, elapsed: 44s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 262/287, 5.9 task/s, elapsed: 44s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 263/287, 5.8 task/s, elapsed: 45s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 264/287, 5.8 task/s, elapsed: 46s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 265/287, 5.7 task/s, elapsed: 47s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 266/287, 5.7 task/s, elapsed: 47s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 267/287, 5.7 task/s, elapsed: 47s, ETA: 4s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 268/287, 5.6 task/s, elapsed: 48s, ETA: 3s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 269/287, 5.6 task/s, elapsed: 48s, ETA: 3s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 270/287, 5.6 task/s, elapsed: 48s, ETA: 3s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 271/287, 5.5 task/s, elapsed: 49s, ETA: 3s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 272/287, 5.5 task/s, elapsed: 50s, ETA: 3s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 273/287, 5.5 task/s, elapsed: 50s, ETA: 3s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 274/287, 5.5 task/s, elapsed: 50s, ETA: 2s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 275/287, 5.5 task/s, elapsed: 50s, ETA: 2s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 276/287, 5.3 task/s, elapsed: 52s, ETA: 2s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 277/287, 5.3 task/s, elapsed: 52s, ETA: 2s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 278/287, 5.3 task/s, elapsed: 52s, ETA: 2s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 279/287, 5.2 task/s, elapsed: 53s, ETA: 2s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 280/287, 5.2 task/s, elapsed: 54s, ETA: 1s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 281/287, 5.2 task/s, elapsed: 54s, ETA: 1s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 282/287, 5.2 task/s, elapsed: 54s, ETA: 1s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 283/287, 5.1 task/s, elapsed: 56s, ETA: 1s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 284/287, 4.7 task/s, elapsed: 60s, ETA: 1s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 285/287, 4.6 task/s, elapsed: 61s, ETA: 0s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>> ] 286/287, 1.3 task/s, elapsed: 214s, ETA: 1s
[>>>>>>>>>>>>>>>>>>>>>>>>>>>>>] 287/287, 1.3 task/s, elapsed: 216s, ETA: 0s
dataset version metric mode internvl-chat-20b
---------------------------- --------- ---------------------------- ------ -------------------
mmlu - naive_average gen 46.35
mmlu_pro - - - -
cmmlu - naive_average gen 47.13
ceval - naive_average gen 48.56
agieval - - - -
GaokaoBench - weighted_average gen 32.28
GPQA_extended - - - -
GPQA_main - - - -
GPQA_diamond - - - -
ARC-c - - - -
truthfulqa - - - -
triviaqa 2121ce score gen 31.47
triviaqa_wiki_1shot - - - -
nq 3dcea1 score gen 13.21
C3 8c358f accuracy gen 76.88
race-high 9a54b6 accuracy gen 72.56
flores_100 - - - -
winogrande b36770 accuracy gen 58.72
hellaswag e42710 accuracy gen 53.69
bbh - naive_average gen 36.32
gsm8k 1d7fe4 accuracy gen 40.71
math 393424 accuracy gen 6.96
TheoremQA 6f0af8 score gen 12.25
MathBench - - - -
openai_humaneval 8e312c humaneval_pass@1 gen 32.32
humaneval_plus - - - -
humanevalx - - - -
sanitized_mbpp a447ff score gen 33.07
mbpp_plus - - - -
mbpp_cn 6fb572 score gen 23.40
leval - - - -
leval_closed - - - -
leval_open - - - -
longbench - - - -
longbench_single-document-qa - - - -
longbench_multi-document-qa - - - -
longbench_summarization - - - -
longbench_few-shot-learning - - - -
longbench_synthetic-tasks - - - -
longbench_code-completion - - - -
teval - - - -
teval_zh - - - -
IFEval 3321a3 Prompt-level-strict-accuracy gen 19.78
IFEval 3321a3 Inst-level-strict-accuracy gen 31.89
IFEval 3321a3 Prompt-level-loose-accuracy gen 22.92
IFEval 3321a3 Inst-level-loose-accuracy gen 35.13
12/05 03:52:22 - OpenCompass - INFO - write summary to /mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B/20241205_033944/summary/summary_20241205_033944.txt
12/05 03:52:22 - OpenCompass - INFO - write csv to /mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B/20241205_033944/summary/summary_20241205_033944.csv
|