diff --git a/ckpts/checkpoints_zero_stage_2/global_step100/debug.err b/ckpts/checkpoints_zero_stage_2/global_step100/debug.err new file mode 100644 index 0000000000000000000000000000000000000000..641181bfc2d068694f209feb54138551fbe5f620 --- /dev/null +++ b/ckpts/checkpoints_zero_stage_2/global_step100/debug.err @@ -0,0 +1,120 @@ +/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:481: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead. + _torch_pytree._register_pytree_node( +/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:338: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead. + _torch_pytree._register_pytree_node( +The following values were not passed to `accelerate launch` and had defaults used instead: + `--num_processes` was set to a value of `0` + `--num_machines` was set to a value of `1` + `--mixed_precision` was set to a value of `'no'` + `--dynamo_backend` was set to a value of `'no'` +To avoid this warning pass in values for each of the problematic parameters or run `accelerate config`. +/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:481: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead. + _torch_pytree._register_pytree_node( +/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:338: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead. + _torch_pytree._register_pytree_node( +/usr/local/lib/python3.10/dist-packages/transformers/utils/generic.py:338: UserWarning: torch.utils._pytree._register_pytree_node is deprecated. Please use torch.utils._pytree.register_pytree_node instead. + _torch_pytree._register_pytree_node( +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/__init__.py:46: UserWarning: apex not installed, gpu_migration will not swap api for this package. + warnings.warn( +2024-05-22:13:34:29,783 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:13:34:38,319 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:13:34:38,320 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:13:34:38,320 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/idc/cronscript/ckpts//hf_ckpt//global_step100'} +[W socket.cpp:464] [c10d] The server socket cannot be initialized on [::]:12345 (errno: 97 - Address family not supported by protocol). +[W socket.cpp:697] [c10d] The client socket cannot be initialized to connect to [localhost]:12345 (errno: 97 - Address family not supported by protocol). +[W socket.cpp:697] [c10d] The client socket cannot be initialized to connect to [localhost]:12345 (errno: 97 - Address family not supported by protocol). +[W socket.cpp:697] [c10d] The client socket cannot be initialized to connect to [localhost]:12345 (errno: 97 - Address family not supported by protocol). +2024-05-22:13:34:40,627 INFO [huggingface.py:164] Using device 'cuda' +============================= HABANA PT BRIDGE CONFIGURATION =========================== + PT_HPU_LAZY_MODE = 1 + PT_RECIPE_CACHE_PATH = + PT_CACHE_FOLDER_DELETE = 0 + PT_HPU_RECIPE_CACHE_CONFIG = + PT_HPU_MAX_COMPOUND_OP_SIZE = 9223372036854775807 + PT_HPU_LAZY_ACC_PAR_MODE = 1 + PT_HPU_ENABLE_REFINE_DYNAMIC_SHAPES = 0 +---------------------------: System Configuration :--------------------------- +Num CPU Cores : 160 +CPU RAM : 1056375308 KB +------------------------------------------------------------------------------ +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 + Downloading readme: 0%| | 0.00/9.00k [00:00