582 lines
164 KiB
Plaintext
582 lines
164 KiB
Plaintext
|
/mnt/petrelfs/wangweiyun/miniconda3/envs/internvl_eval2/lib/python3.10/site-packages/bitsandbytes/cextension.py:34: UserWarning: The installed version of bitsandbytes was compiled without GPU support. 8-bit optimizers, 8-bit multiplication, and GPU quantization are unavailable.
|
||
|
warn("The installed version of bitsandbytes was compiled without GPU support. "
|
||
|
/mnt/petrelfs/wangweiyun/miniconda3/envs/internvl_eval2/lib/python3.10/site-packages/bitsandbytes/libbitsandbytes_cpu.so: undefined symbol: cadam32bit_grad_fp32
|
||
|
model path is /mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B
|
||
|
12/05 03:39:44 - OpenCompass - WARNING - No previous results to reuse!
|
||
|
12/05 03:39:44 - OpenCompass - INFO - Reusing experiements from 20241205_033944
|
||
|
12/05 03:39:44 - OpenCompass - INFO - Current exp folder: /mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B/20241205_033944
|
||
|
12/05 03:39:48 - OpenCompass - INFO - Partitioned into 256 tasks.
|
||
|
[ ] 0/256, elapsed: 0s, ETA:use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26604 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99913_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13280 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100181_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24526 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100186_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14432 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100185_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30532 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100173_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24834 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99968_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25967 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100187_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24407 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100182_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26493 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99942_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=21062 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99920_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31494 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100177_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27535 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99966_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16310 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100171_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=21774 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99898_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12469 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100166_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14638 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99929_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16693 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99888_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27499 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99952_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16708 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100060_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27588 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99988_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25288 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99889_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31890 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100174_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26297 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100180_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16397 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100157_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12164 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100168_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26540 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100176_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29112 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100179_params.py
|
||
|
command torchrun --master_port=23949 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99887_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=18811 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99994_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13418 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100169_params.py
|
||
|
use_backend False use_backend{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29222 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100032_params.py
|
||
|
command torchrun --master_port=18950 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99947_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=21842 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100162_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30688 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99870_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17817 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100149_params.py
|
||
|
use_backenduse_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16554 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100139_params.py
|
||
|
False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=18486 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99855_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=18748 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100178_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28115 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100184_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15313 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100158_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28444 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100151_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29823 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99914_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28431 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99932_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30982 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99863_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22312 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99901_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26987 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100175_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19899 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99983_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19490 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100167_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=23066 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100133_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30647 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100135_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13344 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99979_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20192 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99893_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19336 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100160_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22730 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100132_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14349 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100122_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22606 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100141_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31476 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99916_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12584 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99921_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17514 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100142_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22902 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100170_params.py
|
||
|
command torchrun --master_port=17978 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100172_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17883 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99948_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25101 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99909_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12829 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100119_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14215 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100153_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24184 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99939_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24341 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100147_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12512 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99970_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26047 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100134_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12077 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99874_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14692 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100188_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27755 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100116_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29389 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99877_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=18251 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100127_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30001 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99895_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29486 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99886_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24196 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100138_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17277 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99982_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30313 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100155_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25681 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100164_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20526 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99977_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14619 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100111_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26666 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100144_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20493 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99866_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28853 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99962_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22252 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100101_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24525 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100108_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=23600 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100114_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15879 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100159_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12811 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100107_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17538 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100125_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22218 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100183_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26764 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100150_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28394 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100120_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15424 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100078_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19811 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100148_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30170 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99987_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19203 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99900_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12951 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99928_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29541 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99910_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=21271 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99926_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=23626 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99919_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17356 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99931_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16269 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100140_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24375 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100115_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20257 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99915_params.py
|
||
|
use_backend False use_backend False{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15094 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99985_params.py
|
||
|
command torchrun --master_port=15353 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100082_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17610 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99878_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13371 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99986_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29090 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100123_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=21099 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99892_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29294 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100145_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=21659 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100165_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16912 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100152_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13163 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99865_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16231 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100124_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12932 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100083_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14734 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99958_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31341 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100073_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14328 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99876_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20416 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100113_params.py
|
||
|
use_backend False use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command commandtorchrun --master_port=17135 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99879_params.py
|
||
|
torchrun --master_port=18677 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99924_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12094 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99912_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19355 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100112_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15928 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100104_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12862 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100106_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27787 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99864_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13000 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100090_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28745 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100063_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25264 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100126_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16067 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100146_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27299 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100096_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16906 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100156_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19369 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100110_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26611 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99961_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16376 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99963_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27537 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100088_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=21108 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99941_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17468 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100102_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20674 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99957_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29071 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100163_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15347 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100077_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24037 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100131_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12977 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100072_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13057 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100109_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13267 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99959_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15163 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99972_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25976 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100079_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27184 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99944_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31203 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100089_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27583 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99976_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30842 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100118_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25586 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100121_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19830 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100161_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31081 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100095_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29153 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100075_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=23362 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100081_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29435 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99953_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12824 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100093_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=18395 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99884_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24334 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99881_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20392 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99896_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19381 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99984_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16250 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100080_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20045 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100098_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12230 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99908_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=21004 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100117_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30647 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99872_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19046 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99989_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19156 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100062_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28238 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99956_params.py
|
||
|
use_backenduse_backend False False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16891 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100094_params.py
|
||
|
command torchrun --master_port=15867 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99930_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31662 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100128_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28884 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100071_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30497 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100092_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17261 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99954_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25699 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100129_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30047 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99981_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13601 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100066_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}command torchrun --master_port=26219 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100130_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19258 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100154_params.py
|
||
|
command torchrun --master_port=12233 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99894_params.py
|
||
|
use_backenduse_backend
|
||
|
False False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
{'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22309 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100084_params.py
|
||
|
command torchrun --master_port=23255 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100074_params.py
|
||
|
command torchrun --master_port=12847 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100086_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22258 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99868_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=18497 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99967_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29005 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100076_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=23792 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100091_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16578 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99885_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17748 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99991_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30847 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100068_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24524 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99990_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22131 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100085_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15395 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99995_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12427 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100100_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15801 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100099_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31270 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99940_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29268 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99978_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16793 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100065_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17810 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99993_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15272 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99992_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13425 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99936_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=19139 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100069_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26643 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99973_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27017 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100097_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=20868 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100103_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25388 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99883_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29743 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99975_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31817 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99911_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13689 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99933_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26617 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100070_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22327 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99918_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12446 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99938_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28013 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99969_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13858 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99974_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25635 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99923_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17105 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99880_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31600 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99927_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27439 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100105_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=24107 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99859_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28968 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99903_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=15510 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100064_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=18400 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99945_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27945 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99875_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31191 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99899_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28700 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99971_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22613 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99907_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30183 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100087_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=14671 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99935_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25091 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99922_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22640 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99917_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22182 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99960_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17209 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99867_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13371 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99858_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=27849 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99856_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=23151 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99897_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22044 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99965_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28553 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99950_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=12281 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99980_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16734 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99860_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=31318 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99861_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16303 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99882_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=13460 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99873_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26241 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99869_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=23737 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99857_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=28106 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99906_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29863 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99854_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=29951 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99853_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16872 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99902_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30472 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99964_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=25834 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99871_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=17323 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99949_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=30716 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/100187_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=16512 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99888_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=22260 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99966_params.py
|
||
|
use_backend False {'abbr': 'internvl-chat-20b', 'batch_size': 4, 'max_out_len': 1024, 'model_args': {'device': 'cuda'}, 'path': '/mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B', 'run_cfg': {'num_gpus': 1}, 'type': 'opencompass.models.InternVLChat'}
|
||
|
command torchrun --master_port=26044 --nproc_per_node 1 /mnt/hwfile/wangweiyun/workspace_tcx/opencompass/opencompass/tasks/openicl_infer.py tmp/99898_params.py
|
||
|
[ ] 1/256, 0.0 task/s, elapsed: 205s, ETA: 52230s
[ ] 2/256, 0.0 task/s, elapsed: 209s, ETA: 26539s
[ ] 3/256, 0.0 task/s, elapsed: 219s, ETA: 18510s
[ ] 4/256, 0.0 task/s, elapsed: 223s, ETA: 14078s
[ ] 5/256, 0.0 task/s, elapsed: 225s, ETA: 11290s
[ ] 6/256, 0.0 task/s, elapsed: 226s, ETA: 9431s
[ ] 7/256, 0.0 task/s, elapsed: 231s, ETA: 8229s
[ ] 8/256, 0.0 task/s, elapsed: 235s, ETA: 7277s
[> ] 9/256, 0.0 task/s, elapsed: 237s, ETA: 6506s
[> ] 10/256, 0.0 task/s, elapsed: 240s, ETA: 5909s
[> ] 11/256, 0.0 task/s, elapsed: 241s, ETA: 5366s
[> ] 12/256, 0.0 task/s, elapsed: 242s, ETA: 4913s
[> ] 13/256, 0.1 task/s, elapsed: 244s, ETA: 4569s
[> ] 14/256, 0.1 task/s, elapsed: 246s, ETA: 4259s
[> ] 15/256, 0.1 task/s, elapsed: 250s, ETA: 4013s
[> ] 16/256, 0.1 task/s, elapsed: 251s, ETA: 3760s
[> ] 17/256, 0.1 task/s, elapsed: 251s, ETA: 3534s
[>> ] 18/256, 0.1 task/s, elapsed: 254s, ETA: 3364s
[>> ] 19/256, 0.1 task/s, elapsed: 256s, ETA: 3199s
[>> ] 20/256, 0.1 task/s, elapsed: 259s, ETA: 3056s
[>> ] 21/256, 0.1 task/s, elapsed: 261s, ETA: 2917s
[>> ] 22/256, 0.1 task/s, elapsed: 264s, ETA: 2804s
[>> ] 23/256, 0.1 task/s, elapsed: 264s, ETA: 2676s
[>> ] 24/256, 0.1 task/s, elapsed: 264s, ETA: 2554s
[>> ] 25/256, 0.1 task/s, elapsed: 266s, ETA: 2461s
[>>> ] 26/256, 0.1 task/s, elapsed: 271s, ETA: 2393s
[>>> ] 27/256, 0.1 task/s, elapsed: 276s, ETA: 2337s
[>>> ] 28/256, 0.1 task/s, elapsed: 279s, ETA: 2273s
[>>> ] 29/256, 0.1 task/s, elapsed: 281s, ETA: 2203s
[>>> ] 30/256, 0.1 task/s, elapsed: 282s, ETA: 2123s
[>>> ] 31/256, 0.1 task/s, elapsed: 282s, ETA: 2049s
[>>> ] 32/256, 0.1 task/s, elapsed: 283s, ETA: 1978s
[>>> ] 33/256, 0.1 task/s, elapsed: 284s, ETA: 1916s
[>>> ] 34/256, 0.1 task/s, elapsed: 284s, ETA: 1856s
[>>>> ] 35/256, 0.1 task/s, elapsed: 286s, ETA: 1806s
[>>>> ] 36/256, 0.1 task/s, elapsed: 288s, ETA: 1758s
[>>>> ] 37/256, 0.1 task/s, elapsed: 288s, ETA: 1706s
[>>>> ] 38/256, 0.1 task/s, elapsed: 290s, ETA: 1666s
[>>>> ] 39/256, 0.1 task/s, elapsed: 292s, ETA: 1623s
[>>>> ] 40/256, 0.1 task/s, elapsed: 292s, ETA: 1578s
[>>>> ] 41/256, 0.1 task/s, elapsed: 292s, ETA: 1532s
[>>>> ] 42/256, 0.1 task/s, elapsed: 294s, ETA: 1497s
[>>>>> ] 43/256, 0.1 task/s, elapsed: 294s, ETA: 1456s
[>>>>> ] 44/256, 0.1 task/s, elapsed: 295s, ETA: 1421s
[>>>>> ] 45/256, 0.2 task/s, elapsed: 296s, ETA: 1386s
[>>>>> ] 46/256, 0.2 task/s, elapsed: 296s, ETA: 1353s
[>>>>> ] 47/256, 0.2 task/s, elapsed: 296s, ETA: 1318s
[>>>>> ] 48/256, 0.2 task/s, elapsed: 297s, ETA: 1285s
[>>>>> ] 49/256, 0.2 task/s, elapsed: 298s, ETA: 1259s
[>>>>> ] 50/256, 0.2 task/s, elapsed: 299s, ETA: 1230s
[>>>>> ] 51/256, 0.2 task/s, elapsed: 299s, ETA: 1202s
[>>>>>>
|
||
|
12/05 03:48:35 - OpenCompass - INFO - Partitioned into 287 tasks.
|
||
|
[ ] 0/287, elapsed: 0s, ETA:
[ ] 1/287, 0.1 task/s, elapsed: 18s, ETA: 5192s
[ ] 2/287, 0.1 task/s, elapsed: 21s, ETA: 3005s
[ ] 3/287, 0.1 task/s, elapsed: 21s, ETA: 2001s
[ ] 4/287, 0.2 task/s, elapsed: 21s, ETA: 1496s
[ ] 5/287, 0.2 task/s, elapsed: 21s, ETA: 1201s
[ ] 6/287, 0.3 task/s, elapsed: 22s, ETA: 1009s
[ ] 7/287, 0.3 task/s, elapsed: 25s, ETA: 1001s
[ ] 8/287, 0.3 task/s, elapsed: 25s, ETA: 879s
[> ] 9/287, 0.4 task/s, elapsed: 25s, ETA: 780s
[> ] 10/287, 0.4 task/s, elapsed: 25s, ETA: 700s
[> ] 11/287, 0.4 task/s, elapsed: 25s, ETA: 634s
[> ] 12/287, 0.5 task/s, elapsed: 25s, ETA: 580s
[> ] 13/287, 0.5 task/s, elapsed: 25s, ETA: 534s
[> ] 14/287, 0.6 task/s, elapsed: 25s, ETA: 494s
[> ] 15/287, 0.6 task/s, elapsed: 25s, ETA: 460s
[> ] 16/287, 0.6 task/s, elapsed: 25s, ETA: 429s
[> ] 17/287, 0.7 task/s, elapsed: 25s, ETA: 403s
[> ] 18/287, 0.7 task/s, elapsed: 25s, ETA: 379s
[>> ] 19/287, 0.7 task/s, elapsed: 25s, ETA: 358s
[>> ] 20/287, 0.8 task/s, elapsed: 25s, ETA: 339s
[>> ] 21/287, 0.8 task/s, elapsed: 25s, ETA: 322s
[>> ] 22/287, 0.9 task/s, elapsed: 25s, ETA: 306s
[>> ] 23/287, 0.9 task/s, elapsed: 26s, ETA: 293s
[>> ] 24/287, 0.9 task/s, elapsed: 26s, ETA: 280s
[>> ] 25/287, 0.8 task/s, elapsed: 30s, ETA: 319s
[>> ] 26/287, 0.9 task/s, elapsed: 30s, ETA: 306s
[>> ] 27/287, 0.9 task/s, elapsed: 31s, ETA: 294s
[>>> ] 28/287, 0.9 task/s, elapsed: 31s, ETA: 283s
[>>> ] 29/287, 0.9 task/s, elapsed: 31s, ETA: 272s
[>>> ] 30/287, 1.0 task/s, elapsed: 31s, ETA: 262s
[>>> ] 31/287, 1.0 task/s, elapsed: 31s, ETA: 253s
[>>> ] 32/287, 1.0 task/s, elapsed: 31s, ETA: 244s
[>>> ] 33/287, 1.1 task/s, elapsed: 31s, ETA: 236s
[>>> ] 34/287, 1.1 task/s, elapsed: 31s, ETA: 228s
[>>> ] 35/287, 1.1 task/s, elapsed: 31s, ETA: 221s
[>>> ] 36/287, 1.2 task/s, elapsed: 31s, ETA: 214s
[>>> ] 37/287, 1.2 task/s, elapsed: 31s, ETA: 207s
[>>>> ] 38/287, 1.2 task/s, elapsed: 31s, ETA: 201s
[>>>> ] 39/287, 1.3 task/s, elapsed: 31s, ETA: 195s
[>>>> ] 40/287, 1.3 task/s, elapsed: 31s, ETA: 190s
[>>>> ] 41/287, 1.3 task/s, elapsed: 31s, ETA: 184s
[>>>> ] 42/287, 1.4 task/s, elapsed: 31s, ETA: 179s
[>>>> ] 43/287, 1.4 task/s, elapsed: 31s, ETA: 174s
[>>>> ] 44/287, 1.4 task/s, elapsed: 31s, ETA: 170s
[>>>> ] 45/287, 1.5 task/s, elapsed: 31s, ETA: 166s
[>>>> ] 46/287, 1.5 task/s, elapsed: 31s, ETA: 161s
[>>>>> ] 47/287, 1.5 task/s, elapsed: 31s, ETA: 158s
[>>>>> ] 48/287, 1.6 task/s, elapsed: 31s, ETA: 154s
[>>>>> ] 49/287, 1.6 task/s, elapsed: 31s, ETA: 152s
[>>>>> ] 50/287, 1.6 task/s, elapsed: 31s, ETA: 149s
[>>>>>
|
||
|
dataset version metric mode internvl-chat-20b
|
||
|
---------------------------- --------- ---------------------------- ------ -------------------
|
||
|
mmlu - naive_average gen 46.35
|
||
|
mmlu_pro - - - -
|
||
|
cmmlu - naive_average gen 47.13
|
||
|
ceval - naive_average gen 48.56
|
||
|
agieval - - - -
|
||
|
GaokaoBench - weighted_average gen 32.28
|
||
|
GPQA_extended - - - -
|
||
|
GPQA_main - - - -
|
||
|
GPQA_diamond - - - -
|
||
|
ARC-c - - - -
|
||
|
truthfulqa - - - -
|
||
|
triviaqa 2121ce score gen 31.47
|
||
|
triviaqa_wiki_1shot - - - -
|
||
|
nq 3dcea1 score gen 13.21
|
||
|
C3 8c358f accuracy gen 76.88
|
||
|
race-high 9a54b6 accuracy gen 72.56
|
||
|
flores_100 - - - -
|
||
|
winogrande b36770 accuracy gen 58.72
|
||
|
hellaswag e42710 accuracy gen 53.69
|
||
|
bbh - naive_average gen 36.32
|
||
|
gsm8k 1d7fe4 accuracy gen 40.71
|
||
|
math 393424 accuracy gen 6.96
|
||
|
TheoremQA 6f0af8 score gen 12.25
|
||
|
MathBench - - - -
|
||
|
openai_humaneval 8e312c humaneval_pass@1 gen 32.32
|
||
|
humaneval_plus - - - -
|
||
|
humanevalx - - - -
|
||
|
sanitized_mbpp a447ff score gen 33.07
|
||
|
mbpp_plus - - - -
|
||
|
mbpp_cn 6fb572 score gen 23.40
|
||
|
leval - - - -
|
||
|
leval_closed - - - -
|
||
|
leval_open - - - -
|
||
|
longbench - - - -
|
||
|
longbench_single-document-qa - - - -
|
||
|
longbench_multi-document-qa - - - -
|
||
|
longbench_summarization - - - -
|
||
|
longbench_few-shot-learning - - - -
|
||
|
longbench_synthetic-tasks - - - -
|
||
|
longbench_code-completion - - - -
|
||
|
teval - - - -
|
||
|
teval_zh - - - -
|
||
|
IFEval 3321a3 Prompt-level-strict-accuracy gen 19.78
|
||
|
IFEval 3321a3 Inst-level-strict-accuracy gen 31.89
|
||
|
IFEval 3321a3 Prompt-level-loose-accuracy gen 22.92
|
||
|
IFEval 3321a3 Inst-level-loose-accuracy gen 35.13
|
||
|
12/05 03:52:22 - OpenCompass - INFO - write summary to /mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B/20241205_033944/summary/summary_20241205_033944.txt
|
||
|
12/05 03:52:22 - OpenCompass - INFO - write csv to /mnt/petrelfs/wangweiyun/workspace_cz/InternVL/internvl_chat_dev/share_internvl/InternVL2-2B/20241205_033944/summary/summary_20241205_033944.csv
|