diff --git a/.gitattributes b/.gitattributes index 926089bb3984348bcb1b961c04787c470bd2e8ea..a9e5623074d0ce4ac0ca25babbff02bf4cfd097d 100644 --- a/.gitattributes +++ b/.gitattributes @@ -123,3 +123,5 @@ lm-evaluation-harness/wandb/run-20240608_150638-5wz6ydpy/run-5wz6ydpy.wandb filt lm-evaluation-harness/wandb/run-20240605_123856-sm8adh93/run-sm8adh93.wandb filter=lfs diff=lfs merge=lfs -text lm-evaluation-harness/wandb/run-20240605_062502-fvj55jrc/run-fvj55jrc.wandb filter=lfs diff=lfs merge=lfs -text lm-evaluation-harness/wandb/run-20240608_122925-vmbmpokf/run-vmbmpokf.wandb filter=lfs diff=lfs merge=lfs -text +lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/run-lwgnduuo.wandb filter=lfs diff=lfs merge=lfs -text +lm-evaluation-harness/wandb/run-20240605_140919-mkdnls2x/run-mkdnls2x.wandb filter=lfs diff=lfs merge=lfs -text diff --git a/lm-evaluation-harness/wandb/run-20240514_163424-gpqnyvgo/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_163424-gpqnyvgo/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c256ed8723d7612d6899bd2d6ab07443756a57e4 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163424-gpqnyvgo/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_setup.py:_flush():76] Configure stats pid to 116797 +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 16:34:24,512 WARNING MainThread:116797 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-gpqnyvgo/logs/debug.log +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163424-gpqnyvgo/logs/debug-internal.log +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_init.py:init():560] calling init triggers +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_init.py:init():610] starting backend +2024-05-14 16:34:24,512 INFO MainThread:116797 [wandb_init.py:init():614] setting up manager +2024-05-14 16:34:24,513 INFO MainThread:116797 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 16:34:24,514 INFO MainThread:116797 [wandb_init.py:init():622] backend started and connected +2024-05-14 16:34:24,517 INFO MainThread:116797 [wandb_init.py:init():711] updated telemetry +2024-05-14 16:34:24,529 INFO MainThread:116797 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 16:34:24,803 INFO MainThread:116797 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 16:34:24,883 INFO MainThread:116797 [wandb_run.py:_on_init():2405] got version response +2024-05-14 16:34:24,884 INFO MainThread:116797 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 16:34:25,089 INFO MainThread:116797 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 16:34:25,089 INFO MainThread:116797 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 16:34:25,090 INFO MainThread:116797 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 16:34:25,090 INFO MainThread:116797 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 16:34:25,091 INFO MainThread:116797 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 16:34:57,649 WARNING MsgRouterThr:116797 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..770bd3a3bacd57b25320b34641708cdf1cd2fa1d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715704623 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4d3b688381ec68268e50ff1a4d9306717fd77aaf --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log @@ -0,0 +1,33 @@ + +2024-05-14:16:37:03,996 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:16:37:08,594 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:16:37:08,596 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:16:37:08,597 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step20'} +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/data/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/data/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/data/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 928, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 631, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 686, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 369, in cached_file + raise EnvironmentError( +OSError: /data/cronscript/ckpts//hf_ckpt//global_step20 does not appear to have a file named config.json. Checkout 'https://huggingface.co//data/cronscript/ckpts//hf_ckpt//global_step20/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..1fe4befdcee4f1e8e0b8976d30e1497b840d0933 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T16:37:03.855876", + "startedAt": "2024-05-14T16:37:03.407725", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step20", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3393.177875, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3287.848, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3287.846, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3225.096, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3284.081, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3288.73, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3314.283, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3314.381, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3314.551, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3212.778, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3332.227, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 863.4235038757324 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 5}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b5b86a501a679d5c60f6c7e77a68df56c9b8ca6a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/logs/debug-internal.log @@ -0,0 +1,182 @@ +2024-05-14 16:37:03,420 INFO StreamThr :127606 [internal.py:wandb_internal():85] W&B internal server running at pid: 127606, started at: 2024-05-14 16:37:03.419449 +2024-05-14 16:37:03,422 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: status +2024-05-14 16:37:03,423 INFO WriterThread:127606 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/run-2tgahzpo.wandb +2024-05-14 16:37:03,424 DEBUG SenderThread:127606 [sender.py:send():378] send: header +2024-05-14 16:37:03,432 DEBUG SenderThread:127606 [sender.py:send():378] send: run +2024-05-14 16:37:03,697 INFO SenderThread:127606 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files +2024-05-14 16:37:03,697 INFO SenderThread:127606 [sender.py:_start_run_threads():1123] run started: 2tgahzpo with start time 1715704623.419017 +2024-05-14 16:37:03,704 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 16:37:03,705 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: check_version +2024-05-14 16:37:03,786 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 16:37:03,788 DEBUG HandlerThread:127606 [system_info.py:__init__():26] System info init +2024-05-14 16:37:03,788 DEBUG HandlerThread:127606 [system_info.py:__init__():41] System info init done +2024-05-14 16:37:03,788 INFO HandlerThread:127606 [system_monitor.py:start():194] Starting system monitor +2024-05-14 16:37:03,788 INFO SystemMonitor:127606 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 16:37:03,788 INFO HandlerThread:127606 [system_monitor.py:probe():214] Collecting system info +2024-05-14 16:37:03,789 INFO SystemMonitor:127606 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 16:37:03,789 INFO SystemMonitor:127606 [interfaces.py:start():188] Started disk monitoring +2024-05-14 16:37:03,790 INFO SystemMonitor:127606 [interfaces.py:start():188] Started memory monitoring +2024-05-14 16:37:03,790 INFO SystemMonitor:127606 [interfaces.py:start():188] Started network monitoring +2024-05-14 16:37:03,855 DEBUG HandlerThread:127606 [system_info.py:probe():150] Probing system +2024-05-14 16:37:03,864 DEBUG HandlerThread:127606 [system_info.py:_probe_git():135] Probing git +2024-05-14 16:37:03,887 ERROR HandlerThread:127606 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 16:37:03,887 DEBUG HandlerThread:127606 [system_info.py:_probe_git():143] Probing git done +2024-05-14 16:37:03,887 DEBUG HandlerThread:127606 [system_info.py:probe():198] Probing system done +2024-05-14 16:37:03,887 DEBUG HandlerThread:127606 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T16:37:03.855876', 'startedAt': '2024-05-14T16:37:03.407725', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step20', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3393.177875, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3287.848, 'min': 800.0, 'max': 3400.0}, {'current': 3287.846, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3225.096, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3284.081, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3288.73, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3314.283, 'min': 800.0, 'max': 3400.0}, {'current': 3314.381, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3314.551, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3212.778, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3332.227, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 863.4235038757324}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 16:37:03,888 INFO HandlerThread:127606 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 16:37:03,888 INFO HandlerThread:127606 [system_monitor.py:probe():227] Publishing system info +2024-05-14 16:37:03,896 INFO HandlerThread:127606 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 16:37:03,900 DEBUG SenderThread:127606 [sender.py:send():378] send: files +2024-05-14 16:37:03,900 INFO SenderThread:127606 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 16:37:03,992 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 16:37:03,993 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:37:03,993 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: python_packages +2024-05-14 16:37:03,994 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:37:04,148 DEBUG SenderThread:127606 [sender.py:send():378] send: telemetry +2024-05-14 16:37:04,407 INFO wandb-upload_0:127606 [upload_job.py:push():130] Uploaded file /tmp/tmp4r4tsvwxwandb/jwava83f-wandb-metadata.json +2024-05-14 16:37:04,699 INFO Thread-12 :127606 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/requirements.txt +2024-05-14 16:37:04,699 INFO Thread-12 :127606 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-metadata.json +2024-05-14 16:37:04,699 INFO Thread-12 :127606 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log +2024-05-14 16:37:06,699 INFO Thread-12 :127606 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log +2024-05-14 16:37:08,596 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:37:09,680 DEBUG SenderThread:127606 [sender.py:send():378] send: exit +2024-05-14 16:37:09,681 INFO SenderThread:127606 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 16:37:09,681 INFO SenderThread:127606 [sender.py:send_exit():587] handling runtime: 5 +2024-05-14 16:37:09,682 INFO SenderThread:127606 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:37:09,682 INFO SenderThread:127606 [sender.py:send_exit():593] send defer +2024-05-14 16:37:09,682 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:09,682 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 16:37:09,683 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:09,683 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 16:37:09,683 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 1 +2024-05-14 16:37:09,683 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:09,683 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 16:37:09,683 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:09,683 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 16:37:09,683 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 2 +2024-05-14 16:37:09,683 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:09,683 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 16:37:09,683 INFO HandlerThread:127606 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 16:37:09,683 DEBUG SystemMonitor:127606 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 16:37:09,684 INFO HandlerThread:127606 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 16:37:09,684 DEBUG SystemMonitor:127606 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 16:37:09,684 INFO HandlerThread:127606 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 16:37:09,684 DEBUG SystemMonitor:127606 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 16:37:09,684 INFO HandlerThread:127606 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 16:37:09,685 INFO HandlerThread:127606 [interfaces.py:finish():200] Joined network monitor +2024-05-14 16:37:09,686 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:09,686 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 16:37:09,686 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 3 +2024-05-14 16:37:09,686 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:09,686 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 16:37:09,686 DEBUG SenderThread:127606 [sender.py:send():378] send: stats +2024-05-14 16:37:09,687 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:09,687 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 16:37:09,687 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 4 +2024-05-14 16:37:09,687 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:09,687 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 16:37:09,687 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:09,687 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 16:37:09,687 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 5 +2024-05-14 16:37:09,687 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:09,687 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 16:37:09,687 DEBUG SenderThread:127606 [sender.py:send():378] send: summary +2024-05-14 16:37:09,688 INFO SenderThread:127606 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:37:09,688 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:09,688 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 16:37:09,688 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 6 +2024-05-14 16:37:09,688 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:09,688 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 16:37:09,688 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:09,688 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 16:37:09,691 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:37:09,701 INFO Thread-12 :127606 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-summary.json +2024-05-14 16:37:09,766 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 7 +2024-05-14 16:37:09,766 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:09,766 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 16:37:09,766 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:09,766 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 16:37:10,681 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:37:10,701 INFO Thread-12 :127606 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/config.yaml +2024-05-14 16:37:10,702 INFO Thread-12 :127606 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log +2024-05-14 16:37:12,168 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 8 +2024-05-14 16:37:12,168 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:37:12,168 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:12,168 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 16:37:12,168 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:12,168 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 16:37:12,169 INFO SenderThread:127606 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 16:37:12,169 INFO SenderThread:127606 [job_builder.py:_get_source_type():576] no source found +2024-05-14 16:37:12,169 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 9 +2024-05-14 16:37:12,169 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:12,169 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 16:37:12,169 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:12,169 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 16:37:12,169 INFO SenderThread:127606 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 16:37:12,681 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:37:12,703 INFO SenderThread:127606 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log +2024-05-14 16:37:12,703 INFO SenderThread:127606 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files +2024-05-14 16:37:12,703 INFO SenderThread:127606 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-metadata.json wandb-metadata.json +2024-05-14 16:37:12,703 INFO SenderThread:127606 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/requirements.txt requirements.txt +2024-05-14 16:37:12,703 INFO SenderThread:127606 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-summary.json wandb-summary.json +2024-05-14 16:37:12,704 INFO SenderThread:127606 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/config.yaml config.yaml +2024-05-14 16:37:12,704 INFO SenderThread:127606 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log output.log +2024-05-14 16:37:12,704 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 10 +2024-05-14 16:37:12,704 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:37:12,704 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:12,704 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 16:37:12,708 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:12,708 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 16:37:12,708 INFO SenderThread:127606 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:37:12,934 INFO wandb-upload_0:127606 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/requirements.txt +2024-05-14 16:37:13,125 INFO wandb-upload_1:127606 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/wandb-summary.json +2024-05-14 16:37:13,199 INFO wandb-upload_2:127606 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/config.yaml +2024-05-14 16:37:13,210 INFO wandb-upload_3:127606 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/files/output.log +2024-05-14 16:37:13,410 INFO Thread-11 (_thread_body):127606 [sender.py:transition_state():613] send defer: 11 +2024-05-14 16:37:13,411 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:13,411 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 16:37:13,411 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:13,411 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 16:37:13,411 INFO SenderThread:127606 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 16:37:13,411 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 12 +2024-05-14 16:37:13,411 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:13,411 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 16:37:13,412 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:13,412 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 16:37:13,412 INFO SenderThread:127606 [file_stream.py:finish():601] file stream finish called +2024-05-14 16:37:13,632 INFO SenderThread:127606 [file_stream.py:finish():605] file stream finish is done +2024-05-14 16:37:13,633 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 13 +2024-05-14 16:37:13,633 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:13,633 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 16:37:13,633 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:13,633 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 16:37:13,633 INFO SenderThread:127606 [sender.py:transition_state():613] send defer: 14 +2024-05-14 16:37:13,633 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:37:13,634 DEBUG SenderThread:127606 [sender.py:send():378] send: final +2024-05-14 16:37:13,634 INFO HandlerThread:127606 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 16:37:13,634 DEBUG SenderThread:127606 [sender.py:send():378] send: footer +2024-05-14 16:37:13,634 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: defer +2024-05-14 16:37:13,634 INFO SenderThread:127606 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 16:37:13,634 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:37:13,635 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:37:13,635 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:37:13,635 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 16:37:13,635 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:37:13,635 DEBUG SenderThread:127606 [sender.py:send_request():405] send_request: server_info +2024-05-14 16:37:13,637 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 16:37:13,637 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 16:37:13,637 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 16:37:13,698 INFO MainThread:127606 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 16:37:13,698 INFO MainThread:127606 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 16:37:13,698 INFO MainThread:127606 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 16:37:13,699 DEBUG HandlerThread:127606 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 16:37:13,699 INFO HandlerThread:127606 [handler.py:finish():882] shutting down handler +2024-05-14 16:37:14,635 INFO WriterThread:127606 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/run-2tgahzpo.wandb +2024-05-14 16:37:14,698 INFO SenderThread:127606 [sender.py:finish():1545] shutting down sender +2024-05-14 16:37:14,698 INFO SenderThread:127606 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:37:14,698 INFO SenderThread:127606 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a0a652eece3c91ea82a8bf297f977427f2af11d6 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_setup.py:_flush():76] Configure stats pid to 126396 +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 16:37:03,416 WARNING MainThread:126396 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/logs/debug.log +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/logs/debug-internal.log +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_init.py:init():560] calling init triggers +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_init.py:init():610] starting backend +2024-05-14 16:37:03,416 INFO MainThread:126396 [wandb_init.py:init():614] setting up manager +2024-05-14 16:37:03,418 INFO MainThread:126396 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 16:37:03,418 INFO MainThread:126396 [wandb_init.py:init():622] backend started and connected +2024-05-14 16:37:03,421 INFO MainThread:126396 [wandb_init.py:init():711] updated telemetry +2024-05-14 16:37:03,432 INFO MainThread:126396 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 16:37:03,704 INFO MainThread:126396 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 16:37:03,782 INFO MainThread:126396 [wandb_run.py:_on_init():2405] got version response +2024-05-14 16:37:03,782 INFO MainThread:126396 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 16:37:03,993 INFO MainThread:126396 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 16:37:03,993 INFO MainThread:126396 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 16:37:03,993 INFO MainThread:126396 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 16:37:03,993 INFO MainThread:126396 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 16:37:03,994 INFO MainThread:126396 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 16:37:14,700 WARNING MsgRouterThr:126396 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/run-2tgahzpo.wandb b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/run-2tgahzpo.wandb new file mode 100644 index 0000000000000000000000000000000000000000..5f27e9fb2eb05a84c3eafa732ac23b6c93f44c18 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_163703-2tgahzpo/run-2tgahzpo.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a12ecd60a5a2b60283171ea4a740ee5d5a100250 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716404172 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..41c2a7ecc3d60776a4ca40757c4f4995a785258b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log @@ -0,0 +1,34 @@ + +2024-05-22:18:56:13,442 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:18:56:21,919 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:18:56:21,920 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:18:56:21,920 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step20000'} +2024-05-22:18:56:24,197 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step20000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step20000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..9bdc255809a5e7926c04c34d12a20af07d9efe09 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T18:56:13.233495", + "startedAt": "2024-05-22T18:56:12.704622", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step20000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.5000375, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.63681030273438 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..721716aa69d3f980fdd7c1329d0eb765d62449f4 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-22 18:56:12,728 INFO StreamThr :2582 [internal.py:wandb_internal():85] W&B internal server running at pid: 2582, started at: 2024-05-22 18:56:12.725832 +2024-05-22 18:56:12,733 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: status +2024-05-22 18:56:12,733 INFO WriterThread:2582 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/run-k5jvnks5.wandb +2024-05-22 18:56:12,736 DEBUG SenderThread:2582 [sender.py:send():378] send: header +2024-05-22 18:56:12,739 DEBUG SenderThread:2582 [sender.py:send():378] send: run +2024-05-22 18:56:13,040 INFO SenderThread:2582 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files +2024-05-22 18:56:13,041 INFO SenderThread:2582 [sender.py:_start_run_threads():1123] run started: k5jvnks5 with start time 1716404172.725667 +2024-05-22 18:56:13,044 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 18:56:13,044 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: check_version +2024-05-22 18:56:13,163 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 18:56:13,165 DEBUG HandlerThread:2582 [system_info.py:__init__():26] System info init +2024-05-22 18:56:13,165 DEBUG HandlerThread:2582 [system_info.py:__init__():41] System info init done +2024-05-22 18:56:13,166 INFO HandlerThread:2582 [system_monitor.py:start():194] Starting system monitor +2024-05-22 18:56:13,166 INFO SystemMonitor:2582 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 18:56:13,166 INFO HandlerThread:2582 [system_monitor.py:probe():214] Collecting system info +2024-05-22 18:56:13,173 INFO SystemMonitor:2582 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 18:56:13,173 INFO SystemMonitor:2582 [interfaces.py:start():188] Started disk monitoring +2024-05-22 18:56:13,173 INFO SystemMonitor:2582 [interfaces.py:start():188] Started memory monitoring +2024-05-22 18:56:13,178 INFO SystemMonitor:2582 [interfaces.py:start():188] Started network monitoring +2024-05-22 18:56:13,233 DEBUG HandlerThread:2582 [system_info.py:probe():150] Probing system +2024-05-22 18:56:13,236 DEBUG HandlerThread:2582 [system_info.py:_probe_git():135] Probing git +2024-05-22 18:56:13,246 ERROR HandlerThread:2582 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 18:56:13,247 DEBUG HandlerThread:2582 [system_info.py:_probe_git():143] Probing git done +2024-05-22 18:56:13,247 DEBUG HandlerThread:2582 [system_info.py:probe():198] Probing system done +2024-05-22 18:56:13,247 DEBUG HandlerThread:2582 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T18:56:13.233495', 'startedAt': '2024-05-22T18:56:12.704622', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step20000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.5000375, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.63681030273438}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 18:56:13,247 INFO HandlerThread:2582 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 18:56:13,247 INFO HandlerThread:2582 [system_monitor.py:probe():227] Publishing system info +2024-05-22 18:56:13,250 INFO HandlerThread:2582 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 18:56:13,255 DEBUG SenderThread:2582 [sender.py:send():378] send: files +2024-05-22 18:56:13,255 INFO SenderThread:2582 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 18:56:13,434 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 18:56:13,434 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: python_packages +2024-05-22 18:56:13,435 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 18:56:13,437 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: stop_status +2024-05-22 18:56:13,546 DEBUG SenderThread:2582 [sender.py:send():378] send: telemetry +2024-05-22 18:56:13,858 INFO wandb-upload_0:2582 [upload_job.py:push():130] Uploaded file /tmp/tmpuee1dheowandb/7j67tu7g-wandb-metadata.json +2024-05-22 18:56:14,043 INFO Thread-12 :2582 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log +2024-05-22 18:56:14,043 INFO Thread-12 :2582 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-metadata.json +2024-05-22 18:56:14,043 INFO Thread-12 :2582 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/requirements.txt +2024-05-22 18:56:16,043 INFO Thread-12 :2582 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log +2024-05-22 18:56:18,552 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:56:23,921 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:56:24,050 INFO Thread-12 :2582 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log +2024-05-22 18:56:24,205 DEBUG SenderThread:2582 [sender.py:send():378] send: exit +2024-05-22 18:56:24,205 INFO SenderThread:2582 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 18:56:24,205 INFO SenderThread:2582 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 18:56:24,207 INFO SenderThread:2582 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:56:24,207 INFO SenderThread:2582 [sender.py:send_exit():593] send defer +2024-05-22 18:56:24,207 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:24,207 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 18:56:24,207 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:24,207 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 18:56:24,207 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 1 +2024-05-22 18:56:24,207 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:24,207 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 18:56:24,207 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:24,207 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 18:56:24,208 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 2 +2024-05-22 18:56:24,208 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:24,208 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 18:56:24,208 INFO HandlerThread:2582 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 18:56:24,208 DEBUG SystemMonitor:2582 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 18:56:24,208 DEBUG SystemMonitor:2582 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 18:56:24,208 DEBUG SystemMonitor:2582 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 18:56:24,211 INFO HandlerThread:2582 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 18:56:24,211 INFO HandlerThread:2582 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 18:56:24,211 INFO HandlerThread:2582 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 18:56:24,211 INFO HandlerThread:2582 [interfaces.py:finish():200] Joined network monitor +2024-05-22 18:56:24,211 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:24,211 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 18:56:24,212 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 3 +2024-05-22 18:56:24,212 DEBUG SenderThread:2582 [sender.py:send():378] send: stats +2024-05-22 18:56:24,213 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:24,213 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 18:56:24,213 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:24,213 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 18:56:24,213 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 4 +2024-05-22 18:56:24,213 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:24,213 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 18:56:24,213 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:24,213 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 18:56:24,213 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 5 +2024-05-22 18:56:24,213 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:24,213 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 18:56:24,213 DEBUG SenderThread:2582 [sender.py:send():378] send: summary +2024-05-22 18:56:24,214 INFO SenderThread:2582 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:56:24,214 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:24,215 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 18:56:24,215 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 6 +2024-05-22 18:56:24,215 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:24,215 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 18:56:24,215 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:24,215 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 18:56:24,219 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:56:24,305 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 7 +2024-05-22 18:56:24,305 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:24,305 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 18:56:24,305 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:24,305 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 18:56:25,051 INFO Thread-12 :2582 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/config.yaml +2024-05-22 18:56:25,052 INFO Thread-12 :2582 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-summary.json +2024-05-22 18:56:25,205 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:56:25,570 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 8 +2024-05-22 18:56:25,570 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:56:25,570 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:25,570 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 18:56:25,570 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:25,570 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 18:56:25,571 INFO SenderThread:2582 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 18:56:25,571 INFO SenderThread:2582 [job_builder.py:_get_source_type():576] no source found +2024-05-22 18:56:25,571 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 9 +2024-05-22 18:56:25,571 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:25,571 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 18:56:25,571 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:25,571 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 18:56:25,571 INFO SenderThread:2582 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 18:56:26,053 INFO SenderThread:2582 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log +2024-05-22 18:56:26,053 INFO SenderThread:2582 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files +2024-05-22 18:56:26,054 INFO SenderThread:2582 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/config.yaml config.yaml +2024-05-22 18:56:26,054 INFO SenderThread:2582 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/requirements.txt requirements.txt +2024-05-22 18:56:26,056 INFO SenderThread:2582 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-summary.json wandb-summary.json +2024-05-22 18:56:26,056 INFO SenderThread:2582 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-metadata.json wandb-metadata.json +2024-05-22 18:56:26,057 INFO SenderThread:2582 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log output.log +2024-05-22 18:56:26,057 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 10 +2024-05-22 18:56:26,057 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:26,057 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 18:56:26,059 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:26,059 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 18:56:26,059 INFO SenderThread:2582 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:56:26,205 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:56:26,205 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:56:26,300 INFO wandb-upload_0:2582 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/config.yaml +2024-05-22 18:56:26,656 INFO wandb-upload_1:2582 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/requirements.txt +2024-05-22 18:56:26,676 INFO wandb-upload_2:2582 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/wandb-summary.json +2024-05-22 18:56:26,678 INFO wandb-upload_3:2582 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/files/output.log +2024-05-22 18:56:26,878 INFO Thread-11 (_thread_body):2582 [sender.py:transition_state():613] send defer: 11 +2024-05-22 18:56:26,878 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:26,878 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 18:56:26,879 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:26,879 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 18:56:26,879 INFO SenderThread:2582 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 18:56:26,879 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 12 +2024-05-22 18:56:26,879 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:26,879 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 18:56:26,879 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:26,879 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 18:56:26,879 INFO SenderThread:2582 [file_stream.py:finish():601] file stream finish called +2024-05-22 18:56:26,954 INFO SenderThread:2582 [file_stream.py:finish():605] file stream finish is done +2024-05-22 18:56:26,954 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 13 +2024-05-22 18:56:26,954 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:26,954 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 18:56:26,955 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:26,955 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 18:56:26,955 INFO SenderThread:2582 [sender.py:transition_state():613] send defer: 14 +2024-05-22 18:56:26,955 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:26,955 INFO HandlerThread:2582 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 18:56:26,955 DEBUG SenderThread:2582 [sender.py:send():378] send: final +2024-05-22 18:56:26,955 DEBUG SenderThread:2582 [sender.py:send():378] send: footer +2024-05-22 18:56:26,955 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:26,955 INFO SenderThread:2582 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 18:56:26,956 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:56:26,956 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:56:26,956 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 18:56:26,956 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 18:56:26,956 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 18:56:26,956 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 18:56:26,957 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:56:26,957 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:56:26,957 DEBUG SenderThread:2582 [sender.py:send_request():405] send_request: server_info +2024-05-22 18:56:27,011 INFO MainThread:2582 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 18:56:27,011 INFO MainThread:2582 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 18:56:27,011 INFO MainThread:2582 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-22 18:56:27,011 DEBUG HandlerThread:2582 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 18:56:27,011 INFO HandlerThread:2582 [handler.py:finish():882] shutting down handler +2024-05-22 18:56:27,956 INFO WriterThread:2582 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/run-k5jvnks5.wandb +2024-05-22 18:56:28,011 INFO SenderThread:2582 [sender.py:finish():1545] shutting down sender +2024-05-22 18:56:28,011 INFO SenderThread:2582 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:56:28,011 INFO SenderThread:2582 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..e992a19314d33d70526bbb191d151bb2e73b2fe4 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 18:56:12,718 INFO MainThread:2427 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 18:56:12,718 INFO MainThread:2427 [wandb_setup.py:_flush():76] Configure stats pid to 2427 +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 18:56:12,719 WARNING MainThread:2427 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/logs/debug.log +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/logs/debug-internal.log +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_init.py:init():560] calling init triggers +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_init.py:init():610] starting backend +2024-05-22 18:56:12,719 INFO MainThread:2427 [wandb_init.py:init():614] setting up manager +2024-05-22 18:56:12,724 INFO MainThread:2427 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 18:56:12,725 INFO MainThread:2427 [wandb_init.py:init():622] backend started and connected +2024-05-22 18:56:12,729 INFO MainThread:2427 [wandb_init.py:init():711] updated telemetry +2024-05-22 18:56:12,739 INFO MainThread:2427 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 18:56:13,044 INFO MainThread:2427 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 18:56:13,157 INFO MainThread:2427 [wandb_run.py:_on_init():2405] got version response +2024-05-22 18:56:13,157 INFO MainThread:2427 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 18:56:13,435 INFO MainThread:2427 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 18:56:13,435 INFO MainThread:2427 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 18:56:13,435 INFO MainThread:2427 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 18:56:13,435 INFO MainThread:2427 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 18:56:13,439 INFO MainThread:2427 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 18:56:28,012 WARNING MsgRouterThr:2427 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/run-k5jvnks5.wandb b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/run-k5jvnks5.wandb new file mode 100644 index 0000000000000000000000000000000000000000..500edac17cddad0cdecf1f85b46f2062e2d572df Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_185612-k5jvnks5/run-k5jvnks5.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..312a5ef065d413d479604380f201e16d3eff3f33 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716440403 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..2403e53a75c20b61a9789fe135da8ff8ab8bc5a0 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:05:00:03,979 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:05:00:12,530 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:05:00:12,531 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:05:00:12,532 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000'} +2024-05-23:05:00:14,835 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..51acfa92029a4304d41d53d1e71b6603e6875011 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T05:00:03.767843", + "startedAt": "2024-05-23T05:00:02.995010", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.104575, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3372.532, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3372.562, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3322.462, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 212.18830108642578 + } + }, + "memory": { + "total": 1007.43798828125 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..a9ba7531d50101464066ba1dafca222a48ffca0d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 05:00:03,014 INFO StreamThr :2517 [internal.py:wandb_internal():85] W&B internal server running at pid: 2517, started at: 2024-05-23 05:00:03.012414 +2024-05-23 05:00:03,019 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: status +2024-05-23 05:00:03,019 INFO WriterThread:2517 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/run-19n1pqay.wandb +2024-05-23 05:00:03,021 DEBUG SenderThread:2517 [sender.py:send():378] send: header +2024-05-23 05:00:03,025 DEBUG SenderThread:2517 [sender.py:send():378] send: run +2024-05-23 05:00:03,464 INFO SenderThread:2517 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files +2024-05-23 05:00:03,464 INFO SenderThread:2517 [sender.py:_start_run_threads():1123] run started: 19n1pqay with start time 1716440403.012502 +2024-05-23 05:00:03,468 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 05:00:03,468 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: check_version +2024-05-23 05:00:03,672 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 05:00:03,675 DEBUG HandlerThread:2517 [system_info.py:__init__():26] System info init +2024-05-23 05:00:03,675 DEBUG HandlerThread:2517 [system_info.py:__init__():41] System info init done +2024-05-23 05:00:03,675 INFO HandlerThread:2517 [system_monitor.py:start():194] Starting system monitor +2024-05-23 05:00:03,675 INFO SystemMonitor:2517 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 05:00:03,675 INFO HandlerThread:2517 [system_monitor.py:probe():214] Collecting system info +2024-05-23 05:00:03,682 INFO SystemMonitor:2517 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 05:00:03,682 INFO SystemMonitor:2517 [interfaces.py:start():188] Started disk monitoring +2024-05-23 05:00:03,683 INFO SystemMonitor:2517 [interfaces.py:start():188] Started memory monitoring +2024-05-23 05:00:03,685 INFO SystemMonitor:2517 [interfaces.py:start():188] Started network monitoring +2024-05-23 05:00:03,767 DEBUG HandlerThread:2517 [system_info.py:probe():150] Probing system +2024-05-23 05:00:03,771 DEBUG HandlerThread:2517 [system_info.py:_probe_git():135] Probing git +2024-05-23 05:00:03,781 ERROR HandlerThread:2517 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 05:00:03,781 DEBUG HandlerThread:2517 [system_info.py:_probe_git():143] Probing git done +2024-05-23 05:00:03,781 DEBUG HandlerThread:2517 [system_info.py:probe():198] Probing system done +2024-05-23 05:00:03,781 DEBUG HandlerThread:2517 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T05:00:03.767843', 'startedAt': '2024-05-23T05:00:02.995010', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step16000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.104575, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3372.532, 'min': 800.0, 'max': 3400.0}, {'current': 3372.562, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3322.462, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 212.18830108642578}}, 'memory': {'total': 1007.43798828125}} +2024-05-23 05:00:03,781 INFO HandlerThread:2517 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 05:00:03,781 INFO HandlerThread:2517 [system_monitor.py:probe():227] Publishing system info +2024-05-23 05:00:03,784 INFO HandlerThread:2517 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 05:00:03,789 DEBUG SenderThread:2517 [sender.py:send():378] send: files +2024-05-23 05:00:03,789 INFO SenderThread:2517 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 05:00:03,972 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 05:00:03,972 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: python_packages +2024-05-23 05:00:03,973 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 05:00:03,974 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: stop_status +2024-05-23 05:00:04,103 DEBUG SenderThread:2517 [sender.py:send():378] send: telemetry +2024-05-23 05:00:04,466 INFO Thread-12 :2517 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/requirements.txt +2024-05-23 05:00:04,466 INFO Thread-12 :2517 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-metadata.json +2024-05-23 05:00:04,466 INFO Thread-12 :2517 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log +2024-05-23 05:00:04,546 INFO wandb-upload_0:2517 [upload_job.py:push():130] Uploaded file /tmp/tmph3zkjg_dwandb/v4uxjjob-wandb-metadata.json +2024-05-23 05:00:06,466 INFO Thread-12 :2517 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log +2024-05-23 05:00:08,109 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:00:13,532 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:00:14,474 INFO Thread-12 :2517 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log +2024-05-23 05:00:14,842 DEBUG SenderThread:2517 [sender.py:send():378] send: exit +2024-05-23 05:00:14,842 INFO SenderThread:2517 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 05:00:14,842 INFO SenderThread:2517 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 05:00:14,843 INFO SenderThread:2517 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:00:14,844 INFO SenderThread:2517 [sender.py:send_exit():593] send defer +2024-05-23 05:00:14,844 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:14,844 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 05:00:14,844 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:14,844 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 05:00:14,844 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 1 +2024-05-23 05:00:14,844 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:14,844 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 05:00:14,844 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:14,844 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 05:00:14,844 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 2 +2024-05-23 05:00:14,844 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:14,845 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 05:00:14,845 INFO HandlerThread:2517 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 05:00:14,845 DEBUG SystemMonitor:2517 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 05:00:14,845 DEBUG SystemMonitor:2517 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 05:00:14,845 DEBUG SystemMonitor:2517 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 05:00:14,848 INFO HandlerThread:2517 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 05:00:14,848 INFO HandlerThread:2517 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 05:00:14,848 INFO HandlerThread:2517 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 05:00:14,848 INFO HandlerThread:2517 [interfaces.py:finish():200] Joined network monitor +2024-05-23 05:00:14,848 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:14,848 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 05:00:14,849 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 3 +2024-05-23 05:00:14,849 DEBUG SenderThread:2517 [sender.py:send():378] send: stats +2024-05-23 05:00:14,850 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:14,850 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 05:00:14,850 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:14,850 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 05:00:14,850 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 4 +2024-05-23 05:00:14,850 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:14,850 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 05:00:14,850 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:14,850 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 05:00:14,850 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 5 +2024-05-23 05:00:14,850 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:14,850 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 05:00:14,850 DEBUG SenderThread:2517 [sender.py:send():378] send: summary +2024-05-23 05:00:14,851 INFO SenderThread:2517 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:00:14,851 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:14,852 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 05:00:14,852 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 6 +2024-05-23 05:00:14,852 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:14,852 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 05:00:14,852 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:14,852 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 05:00:14,856 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:00:14,941 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 7 +2024-05-23 05:00:14,941 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:14,941 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 05:00:14,941 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:14,941 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 05:00:15,475 INFO Thread-12 :2517 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/config.yaml +2024-05-23 05:00:15,475 INFO Thread-12 :2517 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-summary.json +2024-05-23 05:00:15,842 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:00:16,131 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 8 +2024-05-23 05:00:16,131 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:00:16,131 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:16,131 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 05:00:16,131 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:16,131 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 05:00:16,131 INFO SenderThread:2517 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 05:00:16,132 INFO SenderThread:2517 [job_builder.py:_get_source_type():576] no source found +2024-05-23 05:00:16,132 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 9 +2024-05-23 05:00:16,132 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:16,132 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 05:00:16,132 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:16,132 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 05:00:16,132 INFO SenderThread:2517 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 05:00:16,476 INFO SenderThread:2517 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log +2024-05-23 05:00:16,477 INFO SenderThread:2517 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files +2024-05-23 05:00:16,477 INFO SenderThread:2517 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-summary.json wandb-summary.json +2024-05-23 05:00:16,477 INFO SenderThread:2517 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log output.log +2024-05-23 05:00:16,479 INFO SenderThread:2517 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-metadata.json wandb-metadata.json +2024-05-23 05:00:16,480 INFO SenderThread:2517 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/config.yaml config.yaml +2024-05-23 05:00:16,480 INFO SenderThread:2517 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/requirements.txt requirements.txt +2024-05-23 05:00:16,480 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 10 +2024-05-23 05:00:16,480 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:16,480 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 05:00:16,480 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:16,480 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 05:00:16,480 INFO SenderThread:2517 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:00:16,772 INFO wandb-upload_0:2517 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/wandb-summary.json +2024-05-23 05:00:16,842 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:00:16,842 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:00:17,069 INFO wandb-upload_1:2517 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/output.log +2024-05-23 05:00:17,101 INFO wandb-upload_3:2517 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/requirements.txt +2024-05-23 05:00:17,104 INFO wandb-upload_2:2517 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/files/config.yaml +2024-05-23 05:00:17,305 INFO Thread-11 (_thread_body):2517 [sender.py:transition_state():613] send defer: 11 +2024-05-23 05:00:17,305 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:17,305 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 05:00:17,305 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:17,305 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 05:00:17,305 INFO SenderThread:2517 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 05:00:17,305 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 12 +2024-05-23 05:00:17,305 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:17,306 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 05:00:17,306 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:17,306 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 05:00:17,306 INFO SenderThread:2517 [file_stream.py:finish():601] file stream finish called +2024-05-23 05:00:17,389 INFO SenderThread:2517 [file_stream.py:finish():605] file stream finish is done +2024-05-23 05:00:17,389 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 13 +2024-05-23 05:00:17,389 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:17,389 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 05:00:17,389 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:17,389 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 05:00:17,389 INFO SenderThread:2517 [sender.py:transition_state():613] send defer: 14 +2024-05-23 05:00:17,389 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:00:17,389 INFO HandlerThread:2517 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 05:00:17,390 DEBUG SenderThread:2517 [sender.py:send():378] send: final +2024-05-23 05:00:17,390 DEBUG SenderThread:2517 [sender.py:send():378] send: footer +2024-05-23 05:00:17,390 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: defer +2024-05-23 05:00:17,390 INFO SenderThread:2517 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 05:00:17,390 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:00:17,391 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:00:17,391 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 05:00:17,391 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 05:00:17,391 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 05:00:17,391 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 05:00:17,391 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:00:17,391 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:00:17,391 DEBUG SenderThread:2517 [sender.py:send_request():405] send_request: server_info +2024-05-23 05:00:17,445 INFO MainThread:2517 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 05:00:17,445 INFO MainThread:2517 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 05:00:17,445 INFO MainThread:2517 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 05:00:17,445 DEBUG HandlerThread:2517 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 05:00:17,446 INFO HandlerThread:2517 [handler.py:finish():882] shutting down handler +2024-05-23 05:00:18,391 INFO WriterThread:2517 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/run-19n1pqay.wandb +2024-05-23 05:00:18,445 INFO SenderThread:2517 [sender.py:finish():1545] shutting down sender +2024-05-23 05:00:18,445 INFO SenderThread:2517 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:00:18,445 INFO SenderThread:2517 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..735b78d08e7de670951730cbc6822ee1f094356d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_setup.py:_flush():76] Configure stats pid to 2356 +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 05:00:03,006 WARNING MainThread:2356 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 05:00:03,006 INFO MainThread:2356 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/logs/debug.log +2024-05-23 05:00:03,007 INFO MainThread:2356 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/logs/debug-internal.log +2024-05-23 05:00:03,007 INFO MainThread:2356 [wandb_init.py:init():560] calling init triggers +2024-05-23 05:00:03,007 INFO MainThread:2356 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 05:00:03,007 INFO MainThread:2356 [wandb_init.py:init():610] starting backend +2024-05-23 05:00:03,007 INFO MainThread:2356 [wandb_init.py:init():614] setting up manager +2024-05-23 05:00:03,011 INFO MainThread:2356 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 05:00:03,012 INFO MainThread:2356 [wandb_init.py:init():622] backend started and connected +2024-05-23 05:00:03,015 INFO MainThread:2356 [wandb_init.py:init():711] updated telemetry +2024-05-23 05:00:03,024 INFO MainThread:2356 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 05:00:03,468 INFO MainThread:2356 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 05:00:03,666 INFO MainThread:2356 [wandb_run.py:_on_init():2405] got version response +2024-05-23 05:00:03,666 INFO MainThread:2356 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 05:00:03,973 INFO MainThread:2356 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 05:00:03,973 INFO MainThread:2356 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 05:00:03,973 INFO MainThread:2356 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 05:00:03,973 INFO MainThread:2356 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 05:00:03,976 INFO MainThread:2356 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 05:00:18,446 WARNING MsgRouterThr:2356 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/run-19n1pqay.wandb b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/run-19n1pqay.wandb new file mode 100644 index 0000000000000000000000000000000000000000..26ec2357b44562791e1e9b5f26f8adba1f3c421d Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_050002-19n1pqay/run-19n1pqay.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..78ec4be1a756e736b6dec205a6fb3e5690aaea50 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716441333 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e32ca1c76b6fef6bdda575186a787cb42fc39532 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 13}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2f2f98c2f06c5e2a02452b1b372bf5ec93a32389 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/logs/debug-internal.log @@ -0,0 +1,181 @@ +2024-05-23 05:15:33,956 INFO StreamThr :9289 [internal.py:wandb_internal():85] W&B internal server running at pid: 9289, started at: 2024-05-23 05:15:33.954915 +2024-05-23 05:15:33,963 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: status +2024-05-23 05:15:33,966 INFO WriterThread:9289 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/run-6h1pki8r.wandb +2024-05-23 05:15:33,967 DEBUG SenderThread:9289 [sender.py:send():378] send: header +2024-05-23 05:15:33,974 DEBUG SenderThread:9289 [sender.py:send():378] send: run +2024-05-23 05:15:34,228 INFO SenderThread:9289 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files +2024-05-23 05:15:34,228 INFO SenderThread:9289 [sender.py:_start_run_threads():1123] run started: 6h1pki8r with start time 1716441333.95534 +2024-05-23 05:15:34,237 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 05:15:34,238 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: check_version +2024-05-23 05:15:34,485 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 05:15:34,488 DEBUG HandlerThread:9289 [system_info.py:__init__():26] System info init +2024-05-23 05:15:34,488 DEBUG HandlerThread:9289 [system_info.py:__init__():41] System info init done +2024-05-23 05:15:34,488 INFO HandlerThread:9289 [system_monitor.py:start():194] Starting system monitor +2024-05-23 05:15:34,488 INFO SystemMonitor:9289 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 05:15:34,488 INFO HandlerThread:9289 [system_monitor.py:probe():214] Collecting system info +2024-05-23 05:15:34,496 INFO SystemMonitor:9289 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 05:15:34,504 INFO SystemMonitor:9289 [interfaces.py:start():188] Started disk monitoring +2024-05-23 05:15:34,505 INFO SystemMonitor:9289 [interfaces.py:start():188] Started memory monitoring +2024-05-23 05:15:34,506 INFO SystemMonitor:9289 [interfaces.py:start():188] Started network monitoring +2024-05-23 05:15:34,569 DEBUG HandlerThread:9289 [system_info.py:probe():150] Probing system +2024-05-23 05:15:34,587 DEBUG HandlerThread:9289 [system_info.py:_probe_git():135] Probing git +2024-05-23 05:15:34,613 ERROR HandlerThread:9289 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 05:15:34,614 DEBUG HandlerThread:9289 [system_info.py:_probe_git():143] Probing git done +2024-05-23 05:15:34,614 DEBUG HandlerThread:9289 [system_info.py:probe():198] Probing system done +2024-05-23 05:15:34,614 DEBUG HandlerThread:9289 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T05:15:34.569242', 'startedAt': '2024-05-23T05:15:33.899392', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step28000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2687.9523, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3041.103, 'min': 800.0, 'max': 3400.0}, {'current': 3041.105, 'min': 800.0, 'max': 3400.0}, {'current': 3035.541, 'min': 800.0, 'max': 3400.0}, {'current': 3035.774, 'min': 800.0, 'max': 3400.0}, {'current': 3020.002, 'min': 800.0, 'max': 3400.0}, {'current': 3020.142, 'min': 800.0, 'max': 3400.0}, {'current': 3020.015, 'min': 800.0, 'max': 3400.0}, {'current': 3016.74, 'min': 800.0, 'max': 3400.0}, {'current': 3020.156, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3016.692, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3022.464, 'min': 800.0, 'max': 3400.0}, {'current': 3023.397, 'min': 800.0, 'max': 3400.0}, {'current': 3015.736, 'min': 800.0, 'max': 3400.0}, {'current': 3020.154, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3017.272, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3041.259, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3020.2, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3020.164, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3036.186, 'min': 800.0, 'max': 3400.0}, {'current': 3035.901, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3016.882, 'min': 800.0, 'max': 3400.0}, {'current': 3068.716, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3058.626, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3053.671, 'min': 800.0, 'max': 3400.0}, {'current': 3053.717, 'min': 800.0, 'max': 3400.0}, {'current': 3052.642, 'min': 800.0, 'max': 3400.0}, {'current': 3052.83, 'min': 800.0, 'max': 3400.0}, {'current': 3056.233, 'min': 800.0, 'max': 3400.0}, {'current': 3020.537, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3053.873, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3045.909, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3041.648, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3023.303, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 212.18523406982422}}, 'memory': {'total': 1007.43798828125}} +2024-05-23 05:15:34,614 INFO HandlerThread:9289 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 05:15:34,614 INFO HandlerThread:9289 [system_monitor.py:probe():227] Publishing system info +2024-05-23 05:15:34,617 INFO HandlerThread:9289 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 05:15:34,623 DEBUG SenderThread:9289 [sender.py:send():378] send: files +2024-05-23 05:15:34,624 INFO SenderThread:9289 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 05:15:34,854 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 05:15:34,854 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 05:15:34,854 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: python_packages +2024-05-23 05:15:34,871 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: stop_status +2024-05-23 05:15:34,963 DEBUG SenderThread:9289 [sender.py:send():378] send: telemetry +2024-05-23 05:15:35,190 INFO wandb-upload_0:9289 [upload_job.py:push():130] Uploaded file /tmp/tmpjsxc4tv4wandb/yw96yvi2-wandb-metadata.json +2024-05-23 05:15:35,231 INFO Thread-12 :9289 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/wandb-metadata.json +2024-05-23 05:15:35,231 INFO Thread-12 :9289 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/output.log +2024-05-23 05:15:35,231 INFO Thread-12 :9289 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/requirements.txt +2024-05-23 05:15:37,230 INFO Thread-12 :9289 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/output.log +2024-05-23 05:15:38,972 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:15:43,973 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:15:47,258 INFO Thread-12 :9289 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/output.log +2024-05-23 05:15:48,267 DEBUG SenderThread:9289 [sender.py:send():378] send: exit +2024-05-23 05:15:48,267 INFO SenderThread:9289 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 05:15:48,267 INFO SenderThread:9289 [sender.py:send_exit():587] handling runtime: 13 +2024-05-23 05:15:48,268 INFO SenderThread:9289 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:15:48,268 INFO SenderThread:9289 [sender.py:send_exit():593] send defer +2024-05-23 05:15:48,269 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,269 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 05:15:48,269 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,269 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 05:15:48,269 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 1 +2024-05-23 05:15:48,269 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,269 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 05:15:48,269 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,269 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 05:15:48,269 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 2 +2024-05-23 05:15:48,269 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,269 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 05:15:48,269 INFO HandlerThread:9289 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 05:15:48,269 DEBUG SystemMonitor:9289 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 05:15:48,269 DEBUG SystemMonitor:9289 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 05:15:48,269 DEBUG SystemMonitor:9289 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 05:15:48,270 INFO HandlerThread:9289 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 05:15:48,270 INFO HandlerThread:9289 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 05:15:48,271 INFO HandlerThread:9289 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 05:15:48,271 INFO HandlerThread:9289 [interfaces.py:finish():200] Joined network monitor +2024-05-23 05:15:48,271 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,271 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 05:15:48,271 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 3 +2024-05-23 05:15:48,271 DEBUG SenderThread:9289 [sender.py:send():378] send: stats +2024-05-23 05:15:48,271 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,271 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 05:15:48,271 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,272 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 05:15:48,272 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 4 +2024-05-23 05:15:48,272 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,272 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 05:15:48,272 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,272 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 05:15:48,272 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 5 +2024-05-23 05:15:48,272 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,272 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 05:15:48,272 DEBUG SenderThread:9289 [sender.py:send():378] send: summary +2024-05-23 05:15:48,273 INFO SenderThread:9289 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:15:48,273 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,273 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 05:15:48,273 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 6 +2024-05-23 05:15:48,273 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,273 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 05:15:48,273 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,273 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 05:15:48,278 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:15:48,345 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 7 +2024-05-23 05:15:48,345 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,345 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 05:15:48,345 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,345 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 05:15:48,998 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 8 +2024-05-23 05:15:48,999 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:48,999 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 05:15:48,999 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:48,999 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 05:15:48,999 INFO SenderThread:9289 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 05:15:49,001 INFO SenderThread:9289 [job_builder.py:_get_source_type():576] no source found +2024-05-23 05:15:49,001 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 9 +2024-05-23 05:15:49,001 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:49,001 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 05:15:49,001 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:49,001 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 05:15:49,001 INFO SenderThread:9289 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 05:15:49,265 INFO SenderThread:9289 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/config.yaml +2024-05-23 05:15:49,266 INFO SenderThread:9289 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/output.log +2024-05-23 05:15:49,266 INFO SenderThread:9289 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/wandb-summary.json +2024-05-23 05:15:49,266 INFO SenderThread:9289 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files +2024-05-23 05:15:49,267 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:15:49,269 INFO SenderThread:9289 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/config.yaml config.yaml +2024-05-23 05:15:49,269 INFO SenderThread:9289 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/wandb-metadata.json wandb-metadata.json +2024-05-23 05:15:49,272 INFO SenderThread:9289 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/wandb-summary.json wandb-summary.json +2024-05-23 05:15:49,272 INFO SenderThread:9289 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/output.log output.log +2024-05-23 05:15:49,272 INFO SenderThread:9289 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/requirements.txt requirements.txt +2024-05-23 05:15:49,272 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 10 +2024-05-23 05:15:49,273 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:15:49,273 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:49,273 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 05:15:49,273 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:49,273 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 05:15:49,273 INFO SenderThread:9289 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:15:49,532 INFO wandb-upload_0:9289 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/config.yaml +2024-05-23 05:15:49,858 INFO wandb-upload_1:9289 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/wandb-summary.json +2024-05-23 05:15:49,921 INFO wandb-upload_3:9289 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/requirements.txt +2024-05-23 05:15:49,928 INFO wandb-upload_2:9289 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/files/output.log +2024-05-23 05:15:50,129 INFO Thread-11 (_thread_body):9289 [sender.py:transition_state():613] send defer: 11 +2024-05-23 05:15:50,129 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:50,129 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 05:15:50,129 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:50,129 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 05:15:50,130 INFO SenderThread:9289 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 05:15:50,130 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 12 +2024-05-23 05:15:50,130 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:50,130 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 05:15:50,130 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:50,130 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 05:15:50,130 INFO SenderThread:9289 [file_stream.py:finish():601] file stream finish called +2024-05-23 05:15:50,197 INFO SenderThread:9289 [file_stream.py:finish():605] file stream finish is done +2024-05-23 05:15:50,197 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 13 +2024-05-23 05:15:50,197 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:50,197 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 05:15:50,198 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:50,198 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 05:15:50,198 INFO SenderThread:9289 [sender.py:transition_state():613] send defer: 14 +2024-05-23 05:15:50,198 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:15:50,198 INFO HandlerThread:9289 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 05:15:50,198 DEBUG SenderThread:9289 [sender.py:send():378] send: final +2024-05-23 05:15:50,198 DEBUG SenderThread:9289 [sender.py:send():378] send: footer +2024-05-23 05:15:50,198 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: defer +2024-05-23 05:15:50,198 INFO SenderThread:9289 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 05:15:50,199 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:15:50,199 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:15:50,199 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 05:15:50,199 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 05:15:50,199 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 05:15:50,199 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 05:15:50,200 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:15:50,200 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:15:50,200 DEBUG SenderThread:9289 [sender.py:send_request():405] send_request: server_info +2024-05-23 05:15:50,269 INFO MainThread:9289 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 05:15:50,269 INFO MainThread:9289 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 05:15:50,269 INFO MainThread:9289 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 05:15:50,270 DEBUG HandlerThread:9289 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 05:15:50,270 INFO HandlerThread:9289 [handler.py:finish():882] shutting down handler +2024-05-23 05:15:51,200 INFO WriterThread:9289 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/run-6h1pki8r.wandb +2024-05-23 05:15:51,270 INFO SenderThread:9289 [sender.py:finish():1545] shutting down sender +2024-05-23 05:15:51,270 INFO SenderThread:9289 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:15:51,270 INFO SenderThread:9289 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..9b40e64609d41f60ef9dc0a7ee481018574d7cf9 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 05:15:33,947 INFO MainThread:9134 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 05:15:33,947 INFO MainThread:9134 [wandb_setup.py:_flush():76] Configure stats pid to 9134 +2024-05-23 05:15:33,947 INFO MainThread:9134 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 05:15:33,947 INFO MainThread:9134 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 05:15:33,947 INFO MainThread:9134 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 05:15:33,947 INFO MainThread:9134 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 05:15:33,947 WARNING MainThread:9134 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 05:15:33,947 INFO MainThread:9134 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 05:15:33,947 INFO MainThread:9134 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 05:15:33,948 INFO MainThread:9134 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/logs/debug.log +2024-05-23 05:15:33,948 INFO MainThread:9134 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051533-6h1pki8r/logs/debug-internal.log +2024-05-23 05:15:33,948 INFO MainThread:9134 [wandb_init.py:init():560] calling init triggers +2024-05-23 05:15:33,948 INFO MainThread:9134 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 05:15:33,948 INFO MainThread:9134 [wandb_init.py:init():610] starting backend +2024-05-23 05:15:33,948 INFO MainThread:9134 [wandb_init.py:init():614] setting up manager +2024-05-23 05:15:33,951 INFO MainThread:9134 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 05:15:33,954 INFO MainThread:9134 [wandb_init.py:init():622] backend started and connected +2024-05-23 05:15:33,958 INFO MainThread:9134 [wandb_init.py:init():711] updated telemetry +2024-05-23 05:15:33,973 INFO MainThread:9134 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 05:15:34,237 INFO MainThread:9134 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 05:15:34,457 INFO MainThread:9134 [wandb_run.py:_on_init():2405] got version response +2024-05-23 05:15:34,458 INFO MainThread:9134 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 05:15:34,855 INFO MainThread:9134 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 05:15:34,855 INFO MainThread:9134 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 05:15:34,856 INFO MainThread:9134 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 05:15:34,856 INFO MainThread:9134 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 05:15:34,873 INFO MainThread:9134 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 05:15:51,280 WARNING MsgRouterThr:9134 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c482c1706eda7371abbd3e33eb8d7183828a9fbe --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716444719 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..54ad9a356ca8283913ce973ced1103b455d3cd03 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:06:12:00,412 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:06:12:09,921 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:06:12:09,922 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:06:12:09,923 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100'} +2024-05-23:06:12:12,237 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..06367409187973281b315de92dba893df054af06 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T06:12:00.201601", + "startedAt": "2024-05-23T06:11:59.689111", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2326.8229625, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.59414291381836 + } + }, + "memory": { + "total": 1007.4379615783691 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..018b4068e81a342faa2a3b0691dee6965106bcc3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 12}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..1785cd0c20a2ee34fc8a5f5bebbd5fc8678baac4 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 06:11:59,714 INFO StreamThr :809 [internal.py:wandb_internal():85] W&B internal server running at pid: 809, started at: 2024-05-23 06:11:59.711054 +2024-05-23 06:11:59,717 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: status +2024-05-23 06:11:59,717 INFO WriterThread:809 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/run-bmtm0vbf.wandb +2024-05-23 06:11:59,720 DEBUG SenderThread:809 [sender.py:send():378] send: header +2024-05-23 06:11:59,724 DEBUG SenderThread:809 [sender.py:send():378] send: run +2024-05-23 06:11:59,986 INFO SenderThread:809 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files +2024-05-23 06:11:59,986 INFO SenderThread:809 [sender.py:_start_run_threads():1123] run started: bmtm0vbf with start time 1716444719.711131 +2024-05-23 06:11:59,989 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 06:11:59,989 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: check_version +2024-05-23 06:12:00,106 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 06:12:00,108 DEBUG HandlerThread:809 [system_info.py:__init__():26] System info init +2024-05-23 06:12:00,108 DEBUG HandlerThread:809 [system_info.py:__init__():41] System info init done +2024-05-23 06:12:00,108 INFO HandlerThread:809 [system_monitor.py:start():194] Starting system monitor +2024-05-23 06:12:00,109 INFO SystemMonitor:809 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 06:12:00,109 INFO HandlerThread:809 [system_monitor.py:probe():214] Collecting system info +2024-05-23 06:12:00,116 INFO SystemMonitor:809 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 06:12:00,116 INFO SystemMonitor:809 [interfaces.py:start():188] Started disk monitoring +2024-05-23 06:12:00,117 INFO SystemMonitor:809 [interfaces.py:start():188] Started memory monitoring +2024-05-23 06:12:00,118 INFO SystemMonitor:809 [interfaces.py:start():188] Started network monitoring +2024-05-23 06:12:00,201 DEBUG HandlerThread:809 [system_info.py:probe():150] Probing system +2024-05-23 06:12:00,204 DEBUG HandlerThread:809 [system_info.py:_probe_git():135] Probing git +2024-05-23 06:12:00,214 ERROR HandlerThread:809 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 06:12:00,214 DEBUG HandlerThread:809 [system_info.py:_probe_git():143] Probing git done +2024-05-23 06:12:00,214 DEBUG HandlerThread:809 [system_info.py:probe():198] Probing system done +2024-05-23 06:12:00,214 DEBUG HandlerThread:809 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T06:12:00.201601', 'startedAt': '2024-05-23T06:11:59.689111', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2326.8229625, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.59414291381836}}, 'memory': {'total': 1007.4379615783691}} +2024-05-23 06:12:00,215 INFO HandlerThread:809 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 06:12:00,215 INFO HandlerThread:809 [system_monitor.py:probe():227] Publishing system info +2024-05-23 06:12:00,218 INFO HandlerThread:809 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 06:12:00,224 DEBUG SenderThread:809 [sender.py:send():378] send: files +2024-05-23 06:12:00,224 INFO SenderThread:809 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 06:12:00,405 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 06:12:00,405 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: python_packages +2024-05-23 06:12:00,408 DEBUG SenderThread:809 [sender.py:send():378] send: telemetry +2024-05-23 06:12:00,408 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 06:12:00,408 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: stop_status +2024-05-23 06:12:00,888 INFO wandb-upload_0:809 [upload_job.py:push():130] Uploaded file /tmp/tmp4l68uaubwandb/c2vw7t3d-wandb-metadata.json +2024-05-23 06:12:00,988 INFO Thread-12 :809 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-metadata.json +2024-05-23 06:12:00,988 INFO Thread-12 :809 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log +2024-05-23 06:12:00,988 INFO Thread-12 :809 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/requirements.txt +2024-05-23 06:12:02,988 INFO Thread-12 :809 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log +2024-05-23 06:12:05,514 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:12:10,923 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:12:10,994 INFO Thread-12 :809 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log +2024-05-23 06:12:12,249 DEBUG SenderThread:809 [sender.py:send():378] send: exit +2024-05-23 06:12:12,249 INFO SenderThread:809 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 06:12:12,249 INFO SenderThread:809 [sender.py:send_exit():587] handling runtime: 12 +2024-05-23 06:12:12,250 INFO SenderThread:809 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 06:12:12,250 INFO SenderThread:809 [sender.py:send_exit():593] send defer +2024-05-23 06:12:12,251 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:12,251 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 06:12:12,251 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:12,251 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 06:12:12,251 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 1 +2024-05-23 06:12:12,251 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:12,251 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 06:12:12,251 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:12,251 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 06:12:12,251 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 2 +2024-05-23 06:12:12,251 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:12,251 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 06:12:12,251 INFO HandlerThread:809 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 06:12:12,251 DEBUG SystemMonitor:809 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 06:12:12,251 DEBUG SystemMonitor:809 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 06:12:12,252 DEBUG SystemMonitor:809 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 06:12:12,254 INFO HandlerThread:809 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 06:12:12,254 INFO HandlerThread:809 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 06:12:12,255 INFO HandlerThread:809 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 06:12:12,255 INFO HandlerThread:809 [interfaces.py:finish():200] Joined network monitor +2024-05-23 06:12:12,255 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:12,255 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 06:12:12,255 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 3 +2024-05-23 06:12:12,255 DEBUG SenderThread:809 [sender.py:send():378] send: stats +2024-05-23 06:12:12,256 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:12,256 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 06:12:12,257 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:12,257 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 06:12:12,257 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 4 +2024-05-23 06:12:12,257 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:12,257 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 06:12:12,257 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:12,257 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 06:12:12,257 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 5 +2024-05-23 06:12:12,257 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:12,257 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 06:12:12,257 DEBUG SenderThread:809 [sender.py:send():378] send: summary +2024-05-23 06:12:12,258 INFO SenderThread:809 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 06:12:12,258 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:12,258 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 06:12:12,258 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 6 +2024-05-23 06:12:12,258 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:12,259 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 06:12:12,259 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:12,259 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 06:12:12,263 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:12:12,343 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 7 +2024-05-23 06:12:12,344 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:12,344 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 06:12:12,344 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:12,344 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 06:12:12,997 INFO Thread-12 :809 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/config.yaml +2024-05-23 06:12:12,997 INFO Thread-12 :809 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-summary.json +2024-05-23 06:12:13,249 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:12:14,533 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 8 +2024-05-23 06:12:14,534 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:12:14,534 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:14,534 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 06:12:14,534 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:14,534 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 06:12:14,534 INFO SenderThread:809 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 06:12:14,535 INFO SenderThread:809 [job_builder.py:_get_source_type():576] no source found +2024-05-23 06:12:14,535 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 9 +2024-05-23 06:12:14,535 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:14,535 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 06:12:14,535 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:14,535 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 06:12:14,535 INFO SenderThread:809 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 06:12:14,999 INFO SenderThread:809 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log +2024-05-23 06:12:15,000 INFO SenderThread:809 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files +2024-05-23 06:12:15,000 INFO SenderThread:809 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/requirements.txt requirements.txt +2024-05-23 06:12:15,000 INFO SenderThread:809 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-summary.json wandb-summary.json +2024-05-23 06:12:15,002 INFO SenderThread:809 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-metadata.json wandb-metadata.json +2024-05-23 06:12:15,004 INFO SenderThread:809 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log output.log +2024-05-23 06:12:15,004 INFO SenderThread:809 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/config.yaml config.yaml +2024-05-23 06:12:15,004 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 10 +2024-05-23 06:12:15,007 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:15,007 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 06:12:15,007 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:15,007 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 06:12:15,007 INFO SenderThread:809 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 06:12:15,249 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:12:15,249 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:12:15,265 INFO wandb-upload_0:809 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/requirements.txt +2024-05-23 06:12:15,620 INFO wandb-upload_2:809 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/output.log +2024-05-23 06:12:15,637 INFO wandb-upload_3:809 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/config.yaml +2024-05-23 06:12:15,643 INFO wandb-upload_1:809 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/files/wandb-summary.json +2024-05-23 06:12:15,843 INFO Thread-11 (_thread_body):809 [sender.py:transition_state():613] send defer: 11 +2024-05-23 06:12:15,843 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:15,843 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 06:12:15,844 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:15,844 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 06:12:15,844 INFO SenderThread:809 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 06:12:15,844 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 12 +2024-05-23 06:12:15,844 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:15,844 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 06:12:15,844 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:15,844 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 06:12:15,844 INFO SenderThread:809 [file_stream.py:finish():601] file stream finish called +2024-05-23 06:12:16,084 INFO SenderThread:809 [file_stream.py:finish():605] file stream finish is done +2024-05-23 06:12:16,084 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 13 +2024-05-23 06:12:16,084 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:16,084 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 06:12:16,084 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:16,084 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 06:12:16,084 INFO SenderThread:809 [sender.py:transition_state():613] send defer: 14 +2024-05-23 06:12:16,084 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:12:16,084 INFO HandlerThread:809 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 06:12:16,085 DEBUG SenderThread:809 [sender.py:send():378] send: final +2024-05-23 06:12:16,085 DEBUG SenderThread:809 [sender.py:send():378] send: footer +2024-05-23 06:12:16,085 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: defer +2024-05-23 06:12:16,085 INFO SenderThread:809 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 06:12:16,085 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:12:16,085 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:12:16,086 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:12:16,086 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 06:12:16,086 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 06:12:16,086 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 06:12:16,086 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 06:12:16,086 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:12:16,086 DEBUG SenderThread:809 [sender.py:send_request():405] send_request: server_info +2024-05-23 06:12:16,141 INFO MainThread:809 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 06:12:16,141 INFO MainThread:809 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 06:12:16,141 INFO MainThread:809 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 06:12:16,141 DEBUG HandlerThread:809 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 06:12:16,142 INFO HandlerThread:809 [handler.py:finish():882] shutting down handler +2024-05-23 06:12:17,086 INFO WriterThread:809 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/run-bmtm0vbf.wandb +2024-05-23 06:12:17,141 INFO SenderThread:809 [sender.py:finish():1545] shutting down sender +2024-05-23 06:12:17,141 INFO SenderThread:809 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 06:12:17,141 INFO SenderThread:809 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..f20f16148a9bf6893f274353fc953c965f7fdf3e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_setup.py:_flush():76] Configure stats pid to 653 +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 06:11:59,705 WARNING MainThread:653 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/logs/debug.log +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/logs/debug-internal.log +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_init.py:init():560] calling init triggers +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_init.py:init():610] starting backend +2024-05-23 06:11:59,705 INFO MainThread:653 [wandb_init.py:init():614] setting up manager +2024-05-23 06:11:59,710 INFO MainThread:653 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 06:11:59,710 INFO MainThread:653 [wandb_init.py:init():622] backend started and connected +2024-05-23 06:11:59,714 INFO MainThread:653 [wandb_init.py:init():711] updated telemetry +2024-05-23 06:11:59,723 INFO MainThread:653 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 06:11:59,989 INFO MainThread:653 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 06:12:00,099 INFO MainThread:653 [wandb_run.py:_on_init():2405] got version response +2024-05-23 06:12:00,100 INFO MainThread:653 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 06:12:00,406 INFO MainThread:653 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 06:12:00,406 INFO MainThread:653 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 06:12:00,407 INFO MainThread:653 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 06:12:00,407 INFO MainThread:653 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 06:12:00,410 INFO MainThread:653 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 06:12:17,142 WARNING MsgRouterThr:653 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/run-bmtm0vbf.wandb b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/run-bmtm0vbf.wandb new file mode 100644 index 0000000000000000000000000000000000000000..22a06a3f92828890ba716b913266c6d838755ff0 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_061159-bmtm0vbf/run-bmtm0vbf.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..b01ff8ed7bb00bf86e0984bf9f6cceb58acef9da --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716450960 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..261039c52e01bcb2ca4f54518b1e62a378e09913 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:07:56:01,511 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:07:56:10,707 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:07:56:10,708 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:07:56:10,708 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000'} +2024-05-23:07:56:13,016 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..bc369d7fa57a1383c51283c31819c7a7e15fb194 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T07:56:01.308767", + "startedAt": "2024-05-23T07:56:00.783146", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2326.875, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.63470077514648 + } + }, + "memory": { + "total": 1007.4379806518555 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b2f82f26f819fa922a18814f82e5054bfd7c8476 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 07:56:00,802 INFO StreamThr :1060 [internal.py:wandb_internal():85] W&B internal server running at pid: 1060, started at: 2024-05-23 07:56:00.801014 +2024-05-23 07:56:00,807 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: status +2024-05-23 07:56:00,809 INFO WriterThread:1060 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/run-j4zxc3ju.wandb +2024-05-23 07:56:00,811 DEBUG SenderThread:1060 [sender.py:send():378] send: header +2024-05-23 07:56:00,814 DEBUG SenderThread:1060 [sender.py:send():378] send: run +2024-05-23 07:56:01,104 INFO SenderThread:1060 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files +2024-05-23 07:56:01,104 INFO SenderThread:1060 [sender.py:_start_run_threads():1123] run started: j4zxc3ju with start time 1716450960.802196 +2024-05-23 07:56:01,110 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 07:56:01,110 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: check_version +2024-05-23 07:56:01,234 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 07:56:01,237 DEBUG HandlerThread:1060 [system_info.py:__init__():26] System info init +2024-05-23 07:56:01,237 DEBUG HandlerThread:1060 [system_info.py:__init__():41] System info init done +2024-05-23 07:56:01,237 INFO HandlerThread:1060 [system_monitor.py:start():194] Starting system monitor +2024-05-23 07:56:01,237 INFO SystemMonitor:1060 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 07:56:01,237 INFO HandlerThread:1060 [system_monitor.py:probe():214] Collecting system info +2024-05-23 07:56:01,244 INFO SystemMonitor:1060 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 07:56:01,244 INFO SystemMonitor:1060 [interfaces.py:start():188] Started disk monitoring +2024-05-23 07:56:01,244 INFO SystemMonitor:1060 [interfaces.py:start():188] Started memory monitoring +2024-05-23 07:56:01,245 INFO SystemMonitor:1060 [interfaces.py:start():188] Started network monitoring +2024-05-23 07:56:01,308 DEBUG HandlerThread:1060 [system_info.py:probe():150] Probing system +2024-05-23 07:56:01,312 DEBUG HandlerThread:1060 [system_info.py:_probe_git():135] Probing git +2024-05-23 07:56:01,322 ERROR HandlerThread:1060 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 07:56:01,322 DEBUG HandlerThread:1060 [system_info.py:_probe_git():143] Probing git done +2024-05-23 07:56:01,322 DEBUG HandlerThread:1060 [system_info.py:probe():198] Probing system done +2024-05-23 07:56:01,322 DEBUG HandlerThread:1060 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T07:56:01.308767', 'startedAt': '2024-05-23T07:56:00.783146', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2326.875, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.63470077514648}}, 'memory': {'total': 1007.4379806518555}} +2024-05-23 07:56:01,322 INFO HandlerThread:1060 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 07:56:01,322 INFO HandlerThread:1060 [system_monitor.py:probe():227] Publishing system info +2024-05-23 07:56:01,325 INFO HandlerThread:1060 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 07:56:01,330 DEBUG SenderThread:1060 [sender.py:send():378] send: files +2024-05-23 07:56:01,330 INFO SenderThread:1060 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 07:56:01,503 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 07:56:01,503 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: python_packages +2024-05-23 07:56:01,504 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 07:56:01,506 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: stop_status +2024-05-23 07:56:01,608 DEBUG SenderThread:1060 [sender.py:send():378] send: telemetry +2024-05-23 07:56:01,906 INFO wandb-upload_0:1060 [upload_job.py:push():130] Uploaded file /tmp/tmpejz80fx7wandb/ev6xjk0d-wandb-metadata.json +2024-05-23 07:56:02,107 INFO Thread-12 :1060 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-metadata.json +2024-05-23 07:56:02,107 INFO Thread-12 :1060 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/requirements.txt +2024-05-23 07:56:02,107 INFO Thread-12 :1060 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log +2024-05-23 07:56:04,107 INFO Thread-12 :1060 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log +2024-05-23 07:56:06,614 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 07:56:11,711 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 07:56:12,113 INFO Thread-12 :1060 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log +2024-05-23 07:56:13,027 DEBUG SenderThread:1060 [sender.py:send():378] send: exit +2024-05-23 07:56:13,028 INFO SenderThread:1060 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 07:56:13,028 INFO SenderThread:1060 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 07:56:13,029 INFO SenderThread:1060 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 07:56:13,029 INFO SenderThread:1060 [sender.py:send_exit():593] send defer +2024-05-23 07:56:13,030 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,030 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 07:56:13,030 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,030 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 07:56:13,030 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 1 +2024-05-23 07:56:13,030 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,030 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 07:56:13,030 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,030 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 07:56:13,030 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 2 +2024-05-23 07:56:13,030 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,030 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 07:56:13,030 INFO HandlerThread:1060 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 07:56:13,030 DEBUG SystemMonitor:1060 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 07:56:13,030 DEBUG SystemMonitor:1060 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 07:56:13,030 DEBUG SystemMonitor:1060 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 07:56:13,031 INFO HandlerThread:1060 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 07:56:13,033 INFO HandlerThread:1060 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 07:56:13,033 INFO HandlerThread:1060 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 07:56:13,034 INFO HandlerThread:1060 [interfaces.py:finish():200] Joined network monitor +2024-05-23 07:56:13,034 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,034 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 07:56:13,034 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 3 +2024-05-23 07:56:13,034 DEBUG SenderThread:1060 [sender.py:send():378] send: stats +2024-05-23 07:56:13,035 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,035 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 07:56:13,035 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,036 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 07:56:13,036 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 4 +2024-05-23 07:56:13,036 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,036 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 07:56:13,036 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,036 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 07:56:13,036 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 5 +2024-05-23 07:56:13,036 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,036 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 07:56:13,036 DEBUG SenderThread:1060 [sender.py:send():378] send: summary +2024-05-23 07:56:13,038 INFO SenderThread:1060 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 07:56:13,038 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,038 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 07:56:13,038 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 6 +2024-05-23 07:56:13,038 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,038 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 07:56:13,038 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,038 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 07:56:13,043 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 07:56:13,114 INFO Thread-12 :1060 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-summary.json +2024-05-23 07:56:13,139 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 7 +2024-05-23 07:56:13,139 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,139 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 07:56:13,139 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,140 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 07:56:13,648 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 8 +2024-05-23 07:56:13,648 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,648 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 07:56:13,648 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,648 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 07:56:13,648 INFO SenderThread:1060 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 07:56:13,649 INFO SenderThread:1060 [job_builder.py:_get_source_type():576] no source found +2024-05-23 07:56:13,649 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 9 +2024-05-23 07:56:13,649 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:13,649 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 07:56:13,649 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:13,649 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 07:56:13,649 INFO SenderThread:1060 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 07:56:14,028 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 07:56:14,116 INFO SenderThread:1060 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/config.yaml +2024-05-23 07:56:14,116 INFO SenderThread:1060 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log +2024-05-23 07:56:14,116 INFO SenderThread:1060 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files +2024-05-23 07:56:14,116 INFO SenderThread:1060 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-summary.json wandb-summary.json +2024-05-23 07:56:14,116 INFO SenderThread:1060 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log output.log +2024-05-23 07:56:14,119 INFO SenderThread:1060 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-metadata.json wandb-metadata.json +2024-05-23 07:56:14,119 INFO SenderThread:1060 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/requirements.txt requirements.txt +2024-05-23 07:56:14,119 INFO SenderThread:1060 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/config.yaml config.yaml +2024-05-23 07:56:14,119 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 10 +2024-05-23 07:56:14,119 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 07:56:14,119 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:14,119 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 07:56:14,120 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:14,120 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 07:56:14,120 INFO SenderThread:1060 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 07:56:14,372 INFO wandb-upload_0:1060 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/wandb-summary.json +2024-05-23 07:56:14,603 INFO wandb-upload_1:1060 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/output.log +2024-05-23 07:56:14,863 INFO wandb-upload_3:1060 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/config.yaml +2024-05-23 07:56:14,865 INFO wandb-upload_2:1060 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/files/requirements.txt +2024-05-23 07:56:15,028 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 07:56:15,028 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 07:56:15,065 INFO Thread-11 (_thread_body):1060 [sender.py:transition_state():613] send defer: 11 +2024-05-23 07:56:15,065 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:15,065 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 07:56:15,065 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:15,065 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 07:56:15,065 INFO SenderThread:1060 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 07:56:15,066 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 12 +2024-05-23 07:56:15,066 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:15,066 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 07:56:15,066 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:15,066 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 07:56:15,066 INFO SenderThread:1060 [file_stream.py:finish():601] file stream finish called +2024-05-23 07:56:15,129 INFO SenderThread:1060 [file_stream.py:finish():605] file stream finish is done +2024-05-23 07:56:15,129 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 13 +2024-05-23 07:56:15,129 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:15,129 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 07:56:15,129 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:15,129 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 07:56:15,129 INFO SenderThread:1060 [sender.py:transition_state():613] send defer: 14 +2024-05-23 07:56:15,129 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: defer +2024-05-23 07:56:15,129 INFO HandlerThread:1060 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 07:56:15,130 DEBUG SenderThread:1060 [sender.py:send():378] send: final +2024-05-23 07:56:15,130 DEBUG SenderThread:1060 [sender.py:send():378] send: footer +2024-05-23 07:56:15,130 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: defer +2024-05-23 07:56:15,130 INFO SenderThread:1060 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 07:56:15,130 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 07:56:15,130 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 07:56:15,131 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 07:56:15,131 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 07:56:15,131 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 07:56:15,131 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 07:56:15,131 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 07:56:15,131 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 07:56:15,131 DEBUG SenderThread:1060 [sender.py:send_request():405] send_request: server_info +2024-05-23 07:56:15,195 INFO MainThread:1060 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 07:56:15,195 INFO MainThread:1060 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 07:56:15,195 INFO MainThread:1060 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 07:56:15,198 DEBUG HandlerThread:1060 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 07:56:15,198 INFO HandlerThread:1060 [handler.py:finish():882] shutting down handler +2024-05-23 07:56:16,131 INFO WriterThread:1060 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/run-j4zxc3ju.wandb +2024-05-23 07:56:16,195 INFO SenderThread:1060 [sender.py:finish():1545] shutting down sender +2024-05-23 07:56:16,195 INFO SenderThread:1060 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 07:56:16,195 INFO SenderThread:1060 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7a91a667279576736991cb509a5ac479aa698b86 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_setup.py:_flush():76] Configure stats pid to 905 +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 07:56:00,796 WARNING MainThread:905 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/logs/debug.log +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/logs/debug-internal.log +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_init.py:init():560] calling init triggers +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_init.py:init():610] starting backend +2024-05-23 07:56:00,796 INFO MainThread:905 [wandb_init.py:init():614] setting up manager +2024-05-23 07:56:00,800 INFO MainThread:905 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 07:56:00,801 INFO MainThread:905 [wandb_init.py:init():622] backend started and connected +2024-05-23 07:56:00,805 INFO MainThread:905 [wandb_init.py:init():711] updated telemetry +2024-05-23 07:56:00,813 INFO MainThread:905 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 07:56:01,110 INFO MainThread:905 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 07:56:01,228 INFO MainThread:905 [wandb_run.py:_on_init():2405] got version response +2024-05-23 07:56:01,228 INFO MainThread:905 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 07:56:01,504 INFO MainThread:905 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 07:56:01,504 INFO MainThread:905 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 07:56:01,505 INFO MainThread:905 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 07:56:01,505 INFO MainThread:905 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 07:56:01,508 INFO MainThread:905 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 07:56:16,198 WARNING MsgRouterThr:905 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/run-j4zxc3ju.wandb b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/run-j4zxc3ju.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ba507414031b87f32eb4b463fd1d0801fa2992cc Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_075600-j4zxc3ju/run-j4zxc3ju.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3ec0f22d45c67c3074c6072fe61cfb18f244d404 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716451429 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1cc8a34ac86bb08b82520cfcbb5b5458030754b5 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:08:03:50,368 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:08:03:58,775 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:08:03:58,776 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:08:03:58,777 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step30000'} +2024-05-23:08:04:01,108 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step30000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step30000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..96381ad5f2de93efadfd91eff3e342f0550a0ffc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T08:03:50.105296", + "startedAt": "2024-05-23T08:03:49.617003", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step30000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.4999875000003, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.6415252685547 + } + }, + "memory": { + "total": 1007.4379806518555 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2e08744d0090e3e368fe7e48f7d2d44aecfc5c9f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 08:03:49,648 INFO StreamThr :3821 [internal.py:wandb_internal():85] W&B internal server running at pid: 3821, started at: 2024-05-23 08:03:49.646300 +2024-05-23 08:03:49,652 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: status +2024-05-23 08:03:49,653 INFO WriterThread:3821 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/run-cyy1i6f7.wandb +2024-05-23 08:03:49,655 DEBUG SenderThread:3821 [sender.py:send():378] send: header +2024-05-23 08:03:49,659 DEBUG SenderThread:3821 [sender.py:send():378] send: run +2024-05-23 08:03:49,912 INFO SenderThread:3821 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files +2024-05-23 08:03:49,912 INFO SenderThread:3821 [sender.py:_start_run_threads():1123] run started: cyy1i6f7 with start time 1716451429.64639 +2024-05-23 08:03:49,913 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 08:03:49,913 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: check_version +2024-05-23 08:03:50,028 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 08:03:50,030 DEBUG HandlerThread:3821 [system_info.py:__init__():26] System info init +2024-05-23 08:03:50,030 DEBUG HandlerThread:3821 [system_info.py:__init__():41] System info init done +2024-05-23 08:03:50,030 INFO HandlerThread:3821 [system_monitor.py:start():194] Starting system monitor +2024-05-23 08:03:50,030 INFO SystemMonitor:3821 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 08:03:50,031 INFO HandlerThread:3821 [system_monitor.py:probe():214] Collecting system info +2024-05-23 08:03:50,037 INFO SystemMonitor:3821 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 08:03:50,038 INFO SystemMonitor:3821 [interfaces.py:start():188] Started disk monitoring +2024-05-23 08:03:50,044 INFO SystemMonitor:3821 [interfaces.py:start():188] Started memory monitoring +2024-05-23 08:03:50,044 INFO SystemMonitor:3821 [interfaces.py:start():188] Started network monitoring +2024-05-23 08:03:50,105 DEBUG HandlerThread:3821 [system_info.py:probe():150] Probing system +2024-05-23 08:03:50,108 DEBUG HandlerThread:3821 [system_info.py:_probe_git():135] Probing git +2024-05-23 08:03:50,117 ERROR HandlerThread:3821 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 08:03:50,117 DEBUG HandlerThread:3821 [system_info.py:_probe_git():143] Probing git done +2024-05-23 08:03:50,117 DEBUG HandlerThread:3821 [system_info.py:probe():198] Probing system done +2024-05-23 08:03:50,117 DEBUG HandlerThread:3821 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T08:03:50.105296', 'startedAt': '2024-05-23T08:03:49.617003', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step30000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.4999875000003, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.6415252685547}}, 'memory': {'total': 1007.4379806518555}} +2024-05-23 08:03:50,118 INFO HandlerThread:3821 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 08:03:50,118 INFO HandlerThread:3821 [system_monitor.py:probe():227] Publishing system info +2024-05-23 08:03:50,120 INFO HandlerThread:3821 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 08:03:50,126 DEBUG SenderThread:3821 [sender.py:send():378] send: files +2024-05-23 08:03:50,126 INFO SenderThread:3821 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 08:03:50,328 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 08:03:50,328 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: python_packages +2024-05-23 08:03:50,345 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 08:03:50,345 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: stop_status +2024-05-23 08:03:50,418 DEBUG SenderThread:3821 [sender.py:send():378] send: telemetry +2024-05-23 08:03:50,667 INFO wandb-upload_0:3821 [upload_job.py:push():130] Uploaded file /tmp/tmpok3_nwr_wandb/87i3gzu2-wandb-metadata.json +2024-05-23 08:03:50,914 INFO Thread-12 :3821 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/requirements.txt +2024-05-23 08:03:50,914 INFO Thread-12 :3821 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log +2024-05-23 08:03:50,914 INFO Thread-12 :3821 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-metadata.json +2024-05-23 08:03:52,914 INFO Thread-12 :3821 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log +2024-05-23 08:03:55,426 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:04:00,777 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:04:00,920 INFO Thread-12 :3821 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log +2024-05-23 08:04:01,115 DEBUG SenderThread:3821 [sender.py:send():378] send: exit +2024-05-23 08:04:01,115 INFO SenderThread:3821 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 08:04:01,115 INFO SenderThread:3821 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 08:04:01,117 INFO SenderThread:3821 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 08:04:01,117 INFO SenderThread:3821 [sender.py:send_exit():593] send defer +2024-05-23 08:04:01,117 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:01,117 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 08:04:01,117 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:01,117 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 08:04:01,117 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 1 +2024-05-23 08:04:01,117 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:01,117 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 08:04:01,118 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:01,118 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 08:04:01,118 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 2 +2024-05-23 08:04:01,118 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:01,118 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 08:04:01,118 INFO HandlerThread:3821 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 08:04:01,118 DEBUG SystemMonitor:3821 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 08:04:01,119 DEBUG SystemMonitor:3821 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 08:04:01,119 INFO HandlerThread:3821 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 08:04:01,119 DEBUG SystemMonitor:3821 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 08:04:01,119 INFO HandlerThread:3821 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 08:04:01,120 INFO HandlerThread:3821 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 08:04:01,120 INFO HandlerThread:3821 [interfaces.py:finish():200] Joined network monitor +2024-05-23 08:04:01,121 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:01,121 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 08:04:01,121 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 3 +2024-05-23 08:04:01,121 DEBUG SenderThread:3821 [sender.py:send():378] send: stats +2024-05-23 08:04:01,122 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:01,122 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 08:04:01,122 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:01,122 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 08:04:01,122 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 4 +2024-05-23 08:04:01,122 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:01,122 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 08:04:01,123 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:01,123 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 08:04:01,123 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 5 +2024-05-23 08:04:01,123 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:01,123 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 08:04:01,123 DEBUG SenderThread:3821 [sender.py:send():378] send: summary +2024-05-23 08:04:01,124 INFO SenderThread:3821 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 08:04:01,124 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:01,124 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 08:04:01,124 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 6 +2024-05-23 08:04:01,124 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:01,124 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 08:04:01,124 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:01,124 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 08:04:01,129 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:04:01,211 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 7 +2024-05-23 08:04:01,212 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:01,212 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 08:04:01,212 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:01,212 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 08:04:01,921 INFO Thread-12 :3821 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/config.yaml +2024-05-23 08:04:01,921 INFO Thread-12 :3821 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-summary.json +2024-05-23 08:04:02,115 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:04:02,447 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 8 +2024-05-23 08:04:02,448 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:04:02,448 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:02,448 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 08:04:02,448 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:02,448 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 08:04:02,448 INFO SenderThread:3821 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 08:04:02,449 INFO SenderThread:3821 [job_builder.py:_get_source_type():576] no source found +2024-05-23 08:04:02,449 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 9 +2024-05-23 08:04:02,449 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:02,449 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 08:04:02,449 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:02,449 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 08:04:02,449 INFO SenderThread:3821 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 08:04:02,922 INFO SenderThread:3821 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log +2024-05-23 08:04:02,923 INFO SenderThread:3821 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files +2024-05-23 08:04:02,923 INFO SenderThread:3821 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/config.yaml config.yaml +2024-05-23 08:04:02,923 INFO SenderThread:3821 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log output.log +2024-05-23 08:04:02,926 INFO SenderThread:3821 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/requirements.txt requirements.txt +2024-05-23 08:04:02,926 INFO SenderThread:3821 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-summary.json wandb-summary.json +2024-05-23 08:04:02,926 INFO SenderThread:3821 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-metadata.json wandb-metadata.json +2024-05-23 08:04:02,926 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 10 +2024-05-23 08:04:02,926 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:02,926 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 08:04:02,928 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:02,928 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 08:04:02,928 INFO SenderThread:3821 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 08:04:03,115 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:04:03,116 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:04:03,209 INFO wandb-upload_0:3821 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/config.yaml +2024-05-23 08:04:03,556 INFO wandb-upload_3:3821 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/wandb-summary.json +2024-05-23 08:04:03,560 INFO wandb-upload_2:3821 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/requirements.txt +2024-05-23 08:04:03,615 INFO wandb-upload_1:3821 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/files/output.log +2024-05-23 08:04:03,815 INFO Thread-11 (_thread_body):3821 [sender.py:transition_state():613] send defer: 11 +2024-05-23 08:04:03,815 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:03,815 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 08:04:03,815 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:03,815 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 08:04:03,816 INFO SenderThread:3821 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 08:04:03,816 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 12 +2024-05-23 08:04:03,816 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:03,816 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 08:04:03,816 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:03,816 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 08:04:03,816 INFO SenderThread:3821 [file_stream.py:finish():601] file stream finish called +2024-05-23 08:04:03,879 INFO SenderThread:3821 [file_stream.py:finish():605] file stream finish is done +2024-05-23 08:04:03,879 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 13 +2024-05-23 08:04:03,879 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:03,879 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 08:04:03,879 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:03,879 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 08:04:03,879 INFO SenderThread:3821 [sender.py:transition_state():613] send defer: 14 +2024-05-23 08:04:03,879 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:04:03,879 INFO HandlerThread:3821 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 08:04:03,879 DEBUG SenderThread:3821 [sender.py:send():378] send: final +2024-05-23 08:04:03,880 DEBUG SenderThread:3821 [sender.py:send():378] send: footer +2024-05-23 08:04:03,880 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: defer +2024-05-23 08:04:03,880 INFO SenderThread:3821 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 08:04:03,880 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:04:03,880 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:04:03,880 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 08:04:03,880 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 08:04:03,881 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 08:04:03,881 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 08:04:03,881 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:04:03,881 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:04:03,881 DEBUG SenderThread:3821 [sender.py:send_request():405] send_request: server_info +2024-05-23 08:04:03,955 INFO MainThread:3821 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 08:04:03,955 INFO MainThread:3821 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 08:04:03,955 INFO MainThread:3821 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 08:04:03,955 DEBUG HandlerThread:3821 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 08:04:03,955 INFO HandlerThread:3821 [handler.py:finish():882] shutting down handler +2024-05-23 08:04:04,881 INFO WriterThread:3821 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/run-cyy1i6f7.wandb +2024-05-23 08:04:04,955 INFO SenderThread:3821 [sender.py:finish():1545] shutting down sender +2024-05-23 08:04:04,955 INFO SenderThread:3821 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 08:04:04,955 INFO SenderThread:3821 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..07dc51bbb1d87bd962da0c25eaf4cd917957459f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 08:03:49,640 INFO MainThread:3666 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 08:03:49,640 INFO MainThread:3666 [wandb_setup.py:_flush():76] Configure stats pid to 3666 +2024-05-23 08:03:49,640 INFO MainThread:3666 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 08:03:49,640 INFO MainThread:3666 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 08:03:49,640 INFO MainThread:3666 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 08:03:49,640 INFO MainThread:3666 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 08:03:49,640 WARNING MainThread:3666 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 08:03:49,641 INFO MainThread:3666 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 08:03:49,641 INFO MainThread:3666 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 08:03:49,641 INFO MainThread:3666 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/logs/debug.log +2024-05-23 08:03:49,641 INFO MainThread:3666 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/logs/debug-internal.log +2024-05-23 08:03:49,641 INFO MainThread:3666 [wandb_init.py:init():560] calling init triggers +2024-05-23 08:03:49,641 INFO MainThread:3666 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 08:03:49,641 INFO MainThread:3666 [wandb_init.py:init():610] starting backend +2024-05-23 08:03:49,641 INFO MainThread:3666 [wandb_init.py:init():614] setting up manager +2024-05-23 08:03:49,645 INFO MainThread:3666 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 08:03:49,646 INFO MainThread:3666 [wandb_init.py:init():622] backend started and connected +2024-05-23 08:03:49,649 INFO MainThread:3666 [wandb_init.py:init():711] updated telemetry +2024-05-23 08:03:49,658 INFO MainThread:3666 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 08:03:49,913 INFO MainThread:3666 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 08:03:50,022 INFO MainThread:3666 [wandb_run.py:_on_init():2405] got version response +2024-05-23 08:03:50,022 INFO MainThread:3666 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 08:03:50,365 INFO MainThread:3666 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 08:03:50,365 INFO MainThread:3666 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 08:03:50,365 INFO MainThread:3666 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 08:03:50,365 INFO MainThread:3666 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 08:03:50,367 INFO MainThread:3666 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 08:04:04,956 WARNING MsgRouterThr:3666 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/run-cyy1i6f7.wandb b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/run-cyy1i6f7.wandb new file mode 100644 index 0000000000000000000000000000000000000000..73574e7a53882622e718142b1ebc3e70dec63dbb Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_080349-cyy1i6f7/run-cyy1i6f7.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..172aec385a738823d94459c00e5408bf6a6f2e9d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716467896 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..a8679b78fbe3c4199c0abfcac34cd354c11fb2e4 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:12:38:17,676 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:12:38:26,273 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:12:38:26,274 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:12:38:26,274 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000'} +2024-05-23:12:38:28,589 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..f675c3016b5332c1acf28f436e0b60adeead9c12 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.0 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..b97490fd007aa2a6dc48931025afe231b7657da8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T12:38:17.465712", + "startedAt": "2024-05-23T12:38:16.945763", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.50000625, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 209.57970428466797 + } + }, + "memory": { + "total": 1007.4379425048828 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d68a726b84f5295ffd1b6ef3ad5eb54b83a3a1b7 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 12:38:16,967 INFO StreamThr :4128 [internal.py:wandb_internal():85] W&B internal server running at pid: 4128, started at: 2024-05-23 12:38:16.964038 +2024-05-23 12:38:16,970 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: status +2024-05-23 12:38:16,971 INFO WriterThread:4128 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/run-qeeslnzj.wandb +2024-05-23 12:38:16,973 DEBUG SenderThread:4128 [sender.py:send():378] send: header +2024-05-23 12:38:16,977 DEBUG SenderThread:4128 [sender.py:send():378] send: run +2024-05-23 12:38:17,249 INFO SenderThread:4128 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files +2024-05-23 12:38:17,249 INFO SenderThread:4128 [sender.py:_start_run_threads():1123] run started: qeeslnzj with start time 1716467896.963893 +2024-05-23 12:38:17,253 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 12:38:17,254 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: check_version +2024-05-23 12:38:17,371 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 12:38:17,373 DEBUG HandlerThread:4128 [system_info.py:__init__():26] System info init +2024-05-23 12:38:17,373 DEBUG HandlerThread:4128 [system_info.py:__init__():41] System info init done +2024-05-23 12:38:17,373 INFO HandlerThread:4128 [system_monitor.py:start():194] Starting system monitor +2024-05-23 12:38:17,374 INFO SystemMonitor:4128 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 12:38:17,374 INFO HandlerThread:4128 [system_monitor.py:probe():214] Collecting system info +2024-05-23 12:38:17,380 INFO SystemMonitor:4128 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 12:38:17,381 INFO SystemMonitor:4128 [interfaces.py:start():188] Started disk monitoring +2024-05-23 12:38:17,386 INFO SystemMonitor:4128 [interfaces.py:start():188] Started memory monitoring +2024-05-23 12:38:17,387 INFO SystemMonitor:4128 [interfaces.py:start():188] Started network monitoring +2024-05-23 12:38:17,465 DEBUG HandlerThread:4128 [system_info.py:probe():150] Probing system +2024-05-23 12:38:17,469 DEBUG HandlerThread:4128 [system_info.py:_probe_git():135] Probing git +2024-05-23 12:38:17,479 ERROR HandlerThread:4128 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 12:38:17,479 DEBUG HandlerThread:4128 [system_info.py:_probe_git():143] Probing git done +2024-05-23 12:38:17,479 DEBUG HandlerThread:4128 [system_info.py:probe():198] Probing system done +2024-05-23 12:38:17,479 DEBUG HandlerThread:4128 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T12:38:17.465712', 'startedAt': '2024-05-23T12:38:16.945763', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.50000625, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 209.57970428466797}}, 'memory': {'total': 1007.4379425048828}} +2024-05-23 12:38:17,479 INFO HandlerThread:4128 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 12:38:17,479 INFO HandlerThread:4128 [system_monitor.py:probe():227] Publishing system info +2024-05-23 12:38:17,482 INFO HandlerThread:4128 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 12:38:17,487 DEBUG SenderThread:4128 [sender.py:send():378] send: files +2024-05-23 12:38:17,487 INFO SenderThread:4128 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 12:38:17,668 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 12:38:17,669 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: python_packages +2024-05-23 12:38:17,669 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 12:38:17,672 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: stop_status +2024-05-23 12:38:17,778 DEBUG SenderThread:4128 [sender.py:send():378] send: telemetry +2024-05-23 12:38:18,095 INFO wandb-upload_0:4128 [upload_job.py:push():130] Uploaded file /tmp/tmpz_40x9zzwandb/t0do1uvt-wandb-metadata.json +2024-05-23 12:38:18,251 INFO Thread-12 :4128 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/requirements.txt +2024-05-23 12:38:18,252 INFO Thread-12 :4128 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log +2024-05-23 12:38:18,252 INFO Thread-12 :4128 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-metadata.json +2024-05-23 12:38:20,251 INFO Thread-12 :4128 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log +2024-05-23 12:38:22,781 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:38:28,258 INFO Thread-12 :4128 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log +2024-05-23 12:38:28,275 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:38:28,596 DEBUG SenderThread:4128 [sender.py:send():378] send: exit +2024-05-23 12:38:28,596 INFO SenderThread:4128 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 12:38:28,596 INFO SenderThread:4128 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 12:38:28,597 INFO SenderThread:4128 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 12:38:28,597 INFO SenderThread:4128 [sender.py:send_exit():593] send defer +2024-05-23 12:38:28,598 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:28,598 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 12:38:28,598 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:28,598 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 12:38:28,598 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 1 +2024-05-23 12:38:28,598 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:28,598 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 12:38:28,598 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:28,598 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 12:38:28,598 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 2 +2024-05-23 12:38:28,598 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:28,598 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 12:38:28,598 INFO HandlerThread:4128 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 12:38:28,599 DEBUG SystemMonitor:4128 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 12:38:28,599 DEBUG SystemMonitor:4128 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 12:38:28,599 DEBUG SystemMonitor:4128 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 12:38:28,602 INFO HandlerThread:4128 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 12:38:28,602 INFO HandlerThread:4128 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 12:38:28,602 INFO HandlerThread:4128 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 12:38:28,602 INFO HandlerThread:4128 [interfaces.py:finish():200] Joined network monitor +2024-05-23 12:38:28,602 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:28,602 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 12:38:28,602 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 3 +2024-05-23 12:38:28,602 DEBUG SenderThread:4128 [sender.py:send():378] send: stats +2024-05-23 12:38:28,603 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:28,603 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 12:38:28,604 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:28,604 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 12:38:28,604 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 4 +2024-05-23 12:38:28,604 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:28,604 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 12:38:28,604 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:28,604 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 12:38:28,604 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 5 +2024-05-23 12:38:28,604 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:28,604 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 12:38:28,604 DEBUG SenderThread:4128 [sender.py:send():378] send: summary +2024-05-23 12:38:28,605 INFO SenderThread:4128 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 12:38:28,605 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:28,605 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 12:38:28,605 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 6 +2024-05-23 12:38:28,605 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:28,606 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 12:38:28,606 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:28,606 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 12:38:28,610 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:38:28,700 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 7 +2024-05-23 12:38:28,700 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:28,701 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 12:38:28,701 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:28,701 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 12:38:29,260 INFO Thread-12 :4128 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/config.yaml +2024-05-23 12:38:29,260 INFO Thread-12 :4128 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-summary.json +2024-05-23 12:38:29,596 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:38:29,801 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 8 +2024-05-23 12:38:29,801 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:38:29,801 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:29,801 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 12:38:29,801 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:29,801 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 12:38:29,801 INFO SenderThread:4128 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 12:38:29,802 INFO SenderThread:4128 [job_builder.py:_get_source_type():576] no source found +2024-05-23 12:38:29,802 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 9 +2024-05-23 12:38:29,802 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:29,802 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 12:38:29,802 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:29,802 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 12:38:29,802 INFO SenderThread:4128 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 12:38:30,261 INFO SenderThread:4128 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log +2024-05-23 12:38:30,262 INFO SenderThread:4128 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files +2024-05-23 12:38:30,262 INFO SenderThread:4128 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-metadata.json wandb-metadata.json +2024-05-23 12:38:30,262 INFO SenderThread:4128 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/requirements.txt requirements.txt +2024-05-23 12:38:30,262 INFO SenderThread:4128 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log output.log +2024-05-23 12:38:30,265 INFO SenderThread:4128 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/config.yaml config.yaml +2024-05-23 12:38:30,267 INFO SenderThread:4128 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-summary.json wandb-summary.json +2024-05-23 12:38:30,269 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 10 +2024-05-23 12:38:30,270 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:30,270 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 12:38:30,270 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:30,272 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 12:38:30,272 INFO SenderThread:4128 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 12:38:30,526 INFO wandb-upload_0:4128 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/requirements.txt +2024-05-23 12:38:30,596 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:38:30,596 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:38:30,825 INFO wandb-upload_1:4128 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/output.log +2024-05-23 12:38:30,841 INFO wandb-upload_3:4128 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/wandb-summary.json +2024-05-23 12:38:30,861 INFO wandb-upload_2:4128 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/files/config.yaml +2024-05-23 12:38:31,061 INFO Thread-11 (_thread_body):4128 [sender.py:transition_state():613] send defer: 11 +2024-05-23 12:38:31,062 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:31,062 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 12:38:31,062 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:31,062 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 12:38:31,062 INFO SenderThread:4128 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 12:38:31,062 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 12 +2024-05-23 12:38:31,062 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:31,062 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 12:38:31,063 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:31,063 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 12:38:31,063 INFO SenderThread:4128 [file_stream.py:finish():601] file stream finish called +2024-05-23 12:38:31,122 INFO SenderThread:4128 [file_stream.py:finish():605] file stream finish is done +2024-05-23 12:38:31,122 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 13 +2024-05-23 12:38:31,122 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:31,122 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 12:38:31,123 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:31,123 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 12:38:31,123 INFO SenderThread:4128 [sender.py:transition_state():613] send defer: 14 +2024-05-23 12:38:31,123 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:38:31,123 INFO HandlerThread:4128 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 12:38:31,123 DEBUG SenderThread:4128 [sender.py:send():378] send: final +2024-05-23 12:38:31,123 DEBUG SenderThread:4128 [sender.py:send():378] send: footer +2024-05-23 12:38:31,123 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: defer +2024-05-23 12:38:31,123 INFO SenderThread:4128 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 12:38:31,124 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:38:31,124 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:38:31,124 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 12:38:31,124 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 12:38:31,124 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 12:38:31,124 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 12:38:31,125 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:38:31,125 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:38:31,125 DEBUG SenderThread:4128 [sender.py:send_request():405] send_request: server_info +2024-05-23 12:38:31,186 INFO MainThread:4128 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 12:38:31,186 INFO MainThread:4128 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 12:38:31,187 INFO MainThread:4128 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 12:38:31,187 DEBUG HandlerThread:4128 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 12:38:31,187 INFO HandlerThread:4128 [handler.py:finish():882] shutting down handler +2024-05-23 12:38:32,125 INFO WriterThread:4128 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/run-qeeslnzj.wandb +2024-05-23 12:38:32,186 INFO SenderThread:4128 [sender.py:finish():1545] shutting down sender +2024-05-23 12:38:32,186 INFO SenderThread:4128 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 12:38:32,186 INFO SenderThread:4128 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..6f54c400cde5974dfb36c5985514baad44756454 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_setup.py:_flush():76] Configure stats pid to 3973 +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 12:38:16,959 WARNING MainThread:3973 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/logs/debug.log +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/logs/debug-internal.log +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_init.py:init():560] calling init triggers +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_init.py:init():610] starting backend +2024-05-23 12:38:16,959 INFO MainThread:3973 [wandb_init.py:init():614] setting up manager +2024-05-23 12:38:16,962 INFO MainThread:3973 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 12:38:16,963 INFO MainThread:3973 [wandb_init.py:init():622] backend started and connected +2024-05-23 12:38:16,967 INFO MainThread:3973 [wandb_init.py:init():711] updated telemetry +2024-05-23 12:38:16,976 INFO MainThread:3973 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 12:38:17,253 INFO MainThread:3973 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 12:38:17,365 INFO MainThread:3973 [wandb_run.py:_on_init():2405] got version response +2024-05-23 12:38:17,365 INFO MainThread:3973 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 12:38:17,670 INFO MainThread:3973 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 12:38:17,670 INFO MainThread:3973 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 12:38:17,670 INFO MainThread:3973 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 12:38:17,670 INFO MainThread:3973 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 12:38:17,673 INFO MainThread:3973 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 12:38:32,188 WARNING MsgRouterThr:3973 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/run-qeeslnzj.wandb b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/run-qeeslnzj.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ee31948b30cf57d31a1766668126fb7dfd007b55 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_123816-qeeslnzj/run-qeeslnzj.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..283137f3bfdf871a93676832c625db97c77c7059 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log @@ -0,0 +1,38 @@ + +2024-05-30:07:04:46,014 INFO [__main__.py:251] Verbosity set to INFO +2024-05-30:07:04:56,160 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'boolq', 'copa', 'mrpc', 'piqa', 'sst2', 'winogrande'] +2024-05-30:07:04:56,162 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-30:07:04:56,162 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step20000', 'tokenizer': '/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k/'} +2024-05-30:07:04:58,507 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 237, in __init__ + self._create_tokenizer( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 630, in _create_tokenizer + self.tokenizer = transformers.AutoTokenizer.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/tokenization_auto.py", line 752, in from_pretrained + config = AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 1082, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 644, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 699, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 360, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k/ does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k//main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..7391d0fb5302364497bd6017486c36f0caae0613 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/requirements.txt @@ -0,0 +1,154 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.2 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.15.2 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.36.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..475bec4fe56b7d76103871d5d04b03e15a7d9299 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-30T07:04:45.799280", + "startedAt": "2024-05-30T07:04:45.313252", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step20000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k/", + "--tasks", + "winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq", + "--batch_size", + "auto", + "--wandb_args", + "project=english-eval,group=exp2,name=global_step20000" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-10-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.406825, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.9840431213379 + } + }, + "memory": { + "total": 1007.4379959106445 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e4da61201766f19c52f50d698a897d6d33c28936 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 39}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..cdd02621b22a6fd2bb81faa2f3267c069633cd63 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/logs/debug-internal.log @@ -0,0 +1,194 @@ +2024-05-30 07:04:45,337 INFO StreamThr :900 [internal.py:wandb_internal():85] W&B internal server running at pid: 900, started at: 2024-05-30 07:04:45.335118 +2024-05-30 07:04:45,341 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status +2024-05-30 07:04:45,342 INFO WriterThread:900 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/run-5kdaiq8h.wandb +2024-05-30 07:04:45,344 DEBUG SenderThread:900 [sender.py:send():378] send: header +2024-05-30 07:04:45,348 DEBUG SenderThread:900 [sender.py:send():378] send: run +2024-05-30 07:04:45,600 INFO SenderThread:900 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files +2024-05-30 07:04:45,600 INFO SenderThread:900 [sender.py:_start_run_threads():1123] run started: 5kdaiq8h with start time 1717052685.335222 +2024-05-30 07:04:45,604 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: check_version +2024-05-30 07:04:45,604 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: check_version +2024-05-30 07:04:45,724 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: run_start +2024-05-30 07:04:45,727 DEBUG HandlerThread:900 [system_info.py:__init__():26] System info init +2024-05-30 07:04:45,727 DEBUG HandlerThread:900 [system_info.py:__init__():41] System info init done +2024-05-30 07:04:45,727 INFO HandlerThread:900 [system_monitor.py:start():194] Starting system monitor +2024-05-30 07:04:45,727 INFO SystemMonitor:900 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-30 07:04:45,727 INFO HandlerThread:900 [system_monitor.py:probe():214] Collecting system info +2024-05-30 07:04:45,734 INFO SystemMonitor:900 [interfaces.py:start():188] Started cpu monitoring +2024-05-30 07:04:45,734 INFO SystemMonitor:900 [interfaces.py:start():188] Started disk monitoring +2024-05-30 07:04:45,735 INFO SystemMonitor:900 [interfaces.py:start():188] Started memory monitoring +2024-05-30 07:04:45,735 INFO SystemMonitor:900 [interfaces.py:start():188] Started network monitoring +2024-05-30 07:04:45,799 DEBUG HandlerThread:900 [system_info.py:probe():150] Probing system +2024-05-30 07:04:45,802 DEBUG HandlerThread:900 [system_info.py:_probe_git():135] Probing git +2024-05-30 07:04:45,813 ERROR HandlerThread:900 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-30 07:04:45,813 DEBUG HandlerThread:900 [system_info.py:_probe_git():143] Probing git done +2024-05-30 07:04:45,813 DEBUG HandlerThread:900 [system_info.py:probe():198] Probing system done +2024-05-30 07:04:45,813 DEBUG HandlerThread:900 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-30T07:04:45.799280', 'startedAt': '2024-05-30T07:04:45.313252', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step20000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k/', '--tasks', 'winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq', '--batch_size', 'auto', '--wandb_args', 'project=english-eval,group=exp2,name=global_step20000'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-10-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.406825, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.9840431213379}}, 'memory': {'total': 1007.4379959106445}} +2024-05-30 07:04:45,813 INFO HandlerThread:900 [system_monitor.py:probe():224] Finished collecting system info +2024-05-30 07:04:45,813 INFO HandlerThread:900 [system_monitor.py:probe():227] Publishing system info +2024-05-30 07:04:45,817 INFO HandlerThread:900 [system_monitor.py:probe():229] Finished publishing system info +2024-05-30 07:04:45,824 DEBUG SenderThread:900 [sender.py:send():378] send: files +2024-05-30 07:04:45,824 INFO SenderThread:900 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-30 07:04:46,006 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: python_packages +2024-05-30 07:04:46,006 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: python_packages +2024-05-30 07:04:46,007 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 07:04:46,010 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: stop_status +2024-05-30 07:04:46,144 DEBUG SenderThread:900 [sender.py:send():378] send: telemetry +2024-05-30 07:04:46,412 INFO wandb-upload_0:900 [upload_job.py:push():130] Uploaded file /tmp/tmpde0zlfz6wandb/v3mvkqjg-wandb-metadata.json +2024-05-30 07:04:46,603 INFO Thread-12 :900 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log +2024-05-30 07:04:46,603 INFO Thread-12 :900 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-metadata.json +2024-05-30 07:04:46,603 INFO Thread-12 :900 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/requirements.txt +2024-05-30 07:04:48,603 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log +2024-05-30 07:04:51,149 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:04:56,150 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:04:58,619 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log +2024-05-30 07:05:00,623 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log +2024-05-30 07:05:01,008 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 07:05:01,008 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: stop_status +2024-05-30 07:05:02,080 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:07,080 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:12,081 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:16,008 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 07:05:16,009 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: stop_status +2024-05-30 07:05:17,176 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:17,667 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/config.yaml +2024-05-30 07:05:23,098 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:24,680 INFO Thread-12 :900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log +2024-05-30 07:05:25,003 DEBUG SenderThread:900 [sender.py:send():378] send: exit +2024-05-30 07:05:25,003 INFO SenderThread:900 [sender.py:send_exit():585] handling exit code: 1 +2024-05-30 07:05:25,004 INFO SenderThread:900 [sender.py:send_exit():587] handling runtime: 39 +2024-05-30 07:05:25,005 INFO SenderThread:900 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-30 07:05:25,005 INFO SenderThread:900 [sender.py:send_exit():593] send defer +2024-05-30 07:05:25,005 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,005 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-30 07:05:25,005 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,005 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-30 07:05:25,005 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 1 +2024-05-30 07:05:25,005 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,006 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-30 07:05:25,006 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,006 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-30 07:05:25,006 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 2 +2024-05-30 07:05:25,006 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,006 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-30 07:05:25,006 INFO HandlerThread:900 [system_monitor.py:finish():203] Stopping system monitor +2024-05-30 07:05:25,007 DEBUG SystemMonitor:900 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-30 07:05:25,007 DEBUG SystemMonitor:900 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-30 07:05:25,007 DEBUG SystemMonitor:900 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-30 07:05:25,009 INFO HandlerThread:900 [interfaces.py:finish():200] Joined cpu monitor +2024-05-30 07:05:25,009 INFO HandlerThread:900 [interfaces.py:finish():200] Joined disk monitor +2024-05-30 07:05:25,009 INFO HandlerThread:900 [interfaces.py:finish():200] Joined memory monitor +2024-05-30 07:05:25,009 INFO HandlerThread:900 [interfaces.py:finish():200] Joined network monitor +2024-05-30 07:05:25,010 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,010 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-30 07:05:25,010 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 3 +2024-05-30 07:05:25,010 DEBUG SenderThread:900 [sender.py:send():378] send: stats +2024-05-30 07:05:25,011 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,011 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-30 07:05:25,011 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,011 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-30 07:05:25,011 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 4 +2024-05-30 07:05:25,011 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,011 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-30 07:05:25,011 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,011 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-30 07:05:25,012 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 5 +2024-05-30 07:05:25,012 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,012 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-30 07:05:25,012 DEBUG SenderThread:900 [sender.py:send():378] send: summary +2024-05-30 07:05:25,013 INFO SenderThread:900 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-30 07:05:25,013 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,013 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-30 07:05:25,013 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 6 +2024-05-30 07:05:25,013 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,013 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-30 07:05:25,013 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,013 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-30 07:05:25,013 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 7 +2024-05-30 07:05:25,013 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 07:05:25,013 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:25,013 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-30 07:05:25,014 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:25,014 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-30 07:05:25,681 INFO Thread-12 :900 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-summary.json +2024-05-30 07:05:26,004 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 07:05:26,319 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 8 +2024-05-30 07:05:26,319 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 07:05:26,319 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:26,319 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-30 07:05:26,319 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:26,320 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-30 07:05:26,320 INFO SenderThread:900 [job_builder.py:build():432] Attempting to build job artifact +2024-05-30 07:05:26,320 INFO SenderThread:900 [job_builder.py:_get_source_type():576] no source found +2024-05-30 07:05:26,321 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 9 +2024-05-30 07:05:26,321 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:26,321 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-30 07:05:26,321 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:26,321 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-30 07:05:26,321 INFO SenderThread:900 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-30 07:05:26,683 INFO SenderThread:900 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log +2024-05-30 07:05:26,683 INFO SenderThread:900 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files +2024-05-30 07:05:26,683 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log output.log +2024-05-30 07:05:26,683 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/requirements.txt requirements.txt +2024-05-30 07:05:26,686 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-metadata.json wandb-metadata.json +2024-05-30 07:05:26,686 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-summary.json wandb-summary.json +2024-05-30 07:05:26,686 INFO SenderThread:900 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/config.yaml config.yaml +2024-05-30 07:05:26,686 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 10 +2024-05-30 07:05:26,687 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:26,687 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-30 07:05:26,687 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:26,687 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-30 07:05:26,687 INFO SenderThread:900 [file_pusher.py:finish():169] shutting down file pusher +2024-05-30 07:05:27,004 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 07:05:27,004 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 07:05:27,033 INFO wandb-upload_0:900 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/output.log +2024-05-30 07:05:27,645 INFO wandb-upload_1:900 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/requirements.txt +2024-05-30 07:05:27,646 INFO wandb-upload_2:900 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/wandb-summary.json +2024-05-30 07:05:27,653 INFO wandb-upload_3:900 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/files/config.yaml +2024-05-30 07:05:27,853 INFO Thread-11 (_thread_body):900 [sender.py:transition_state():613] send defer: 11 +2024-05-30 07:05:27,853 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:27,853 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-30 07:05:27,853 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:27,853 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-30 07:05:27,854 INFO SenderThread:900 [file_pusher.py:join():175] waiting for file pusher +2024-05-30 07:05:27,854 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 12 +2024-05-30 07:05:27,854 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:27,854 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-30 07:05:27,854 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:27,854 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-30 07:05:27,854 INFO SenderThread:900 [file_stream.py:finish():601] file stream finish called +2024-05-30 07:05:27,934 INFO SenderThread:900 [file_stream.py:finish():605] file stream finish is done +2024-05-30 07:05:27,934 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 13 +2024-05-30 07:05:27,934 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:27,934 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-30 07:05:27,934 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:27,934 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-30 07:05:27,934 INFO SenderThread:900 [sender.py:transition_state():613] send defer: 14 +2024-05-30 07:05:27,934 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: defer +2024-05-30 07:05:27,934 INFO HandlerThread:900 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-30 07:05:27,935 DEBUG SenderThread:900 [sender.py:send():378] send: final +2024-05-30 07:05:27,935 DEBUG SenderThread:900 [sender.py:send():378] send: footer +2024-05-30 07:05:27,935 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: defer +2024-05-30 07:05:27,935 INFO SenderThread:900 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-30 07:05:27,935 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 07:05:27,935 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 07:05:27,936 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 07:05:27,936 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: server_info +2024-05-30 07:05:27,936 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: get_summary +2024-05-30 07:05:27,936 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-30 07:05:27,936 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-30 07:05:27,936 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 07:05:27,936 DEBUG SenderThread:900 [sender.py:send_request():405] send_request: server_info +2024-05-30 07:05:27,990 INFO MainThread:900 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-30 07:05:27,990 INFO MainThread:900 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-30 07:05:27,990 INFO MainThread:900 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-30 07:05:27,991 DEBUG HandlerThread:900 [handler.py:handle_request():158] handle_request: shutdown +2024-05-30 07:05:27,991 INFO HandlerThread:900 [handler.py:finish():882] shutting down handler +2024-05-30 07:05:28,936 INFO WriterThread:900 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/run-5kdaiq8h.wandb +2024-05-30 07:05:28,990 INFO SenderThread:900 [sender.py:finish():1545] shutting down sender +2024-05-30 07:05:28,990 INFO SenderThread:900 [file_pusher.py:finish():169] shutting down file pusher +2024-05-30 07:05:28,990 INFO SenderThread:900 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/logs/debug.log b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..bc2fbf87ea3bf180b9219db94938c895b940036c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_setup.py:_flush():76] Configure stats pid to 744 +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-30 07:04:45,329 WARNING MainThread:744 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/logs/debug.log +2024-05-30 07:04:45,329 INFO MainThread:744 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/logs/debug-internal.log +2024-05-30 07:04:45,330 INFO MainThread:744 [wandb_init.py:init():560] calling init triggers +2024-05-30 07:04:45,330 INFO MainThread:744 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-30 07:04:45,330 INFO MainThread:744 [wandb_init.py:init():610] starting backend +2024-05-30 07:04:45,330 INFO MainThread:744 [wandb_init.py:init():614] setting up manager +2024-05-30 07:04:45,334 INFO MainThread:744 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-30 07:04:45,335 INFO MainThread:744 [wandb_init.py:init():622] backend started and connected +2024-05-30 07:04:45,338 INFO MainThread:744 [wandb_init.py:init():711] updated telemetry +2024-05-30 07:04:45,347 INFO MainThread:744 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-30 07:04:45,604 INFO MainThread:744 [wandb_run.py:_on_init():2396] communicating current version +2024-05-30 07:04:45,718 INFO MainThread:744 [wandb_run.py:_on_init():2405] got version response +2024-05-30 07:04:45,719 INFO MainThread:744 [wandb_init.py:init():795] starting run threads in backend +2024-05-30 07:04:46,008 INFO MainThread:744 [wandb_run.py:_console_start():2374] atexit reg +2024-05-30 07:04:46,008 INFO MainThread:744 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-30 07:04:46,008 INFO MainThread:744 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-30 07:04:46,008 INFO MainThread:744 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-30 07:04:46,011 INFO MainThread:744 [wandb_init.py:init():838] run started, returning control to user process +2024-05-30 07:05:28,991 WARNING MsgRouterThr:744 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/run-5kdaiq8h.wandb b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/run-5kdaiq8h.wandb new file mode 100644 index 0000000000000000000000000000000000000000..87be12da78da534d5d253d936408adcbce1475ec Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240530_070445-5kdaiq8h/run-5kdaiq8h.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/run-lwgnduuo.wandb b/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/run-lwgnduuo.wandb new file mode 100644 index 0000000000000000000000000000000000000000..5b5884e637f840cf71169af530ce84f91567b9f9 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/run-lwgnduuo.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e150563330859202dd64f9e4e077cfc9b3c6452997901abf7fc8d936aedf8333 +size 1438434 diff --git a/lm-evaluation-harness/wandb/run-20240605_140919-mkdnls2x/run-mkdnls2x.wandb b/lm-evaluation-harness/wandb/run-20240605_140919-mkdnls2x/run-mkdnls2x.wandb new file mode 100644 index 0000000000000000000000000000000000000000..cbc5cb618c6aa1920a2f0661e06ccdda752a69e3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240605_140919-mkdnls2x/run-mkdnls2x.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d79e9e4c6a1dd620ccccda9b4d1672003434b002a23480900ba16f5548bde3b0 +size 1412101 diff --git a/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/config.yaml b/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..369ac92751de613f085525eaf3f287c69c4e9675 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/config.yaml @@ -0,0 +1,375 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.36.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1717649487 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 2 + - 13 + - 23 + - 62 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.36.2 + 8: + - 5 + 13: linux-x86_64 +task_configs: + desc: null + value: + arc_easy: + task: arc_easy + group: + - ai2_arc + dataset_path: allenai/ai2_arc + dataset_name: ARC-Easy + training_split: train + validation_split: validation + test_split: test + doc_to_text: 'Question: {{question}} + + Answer:' + doc_to_target: '{{choices.label.index(answerKey)}}' + doc_to_choice: '{{choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{question}} + + Answer:' + metadata: + version: 1.0 + boolq: + task: boolq + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: boolq + training_split: train + validation_split: validation + doc_to_text: '{{passage}} + + Question: {{question}}? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: passage + metadata: + version: 2.0 + copa: + task: copa + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: copa + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n # Drop the period\n connector =\ + \ {\n \"cause\": \"because\",\n \"effect\": \"therefore\",\n\ + \ }[doc[\"question\"]]\n return doc[\"premise\"].strip()[:-1] + f\"\ + \ {connector}\"\n" + doc_to_target: "def doc_to_target(doc):\n correct_choice = doc[\"choice1\"\ + ] if doc[\"label\"] == 0 else doc[\"choice2\"]\n # Connect the sentences\n\ + \ return \" \" + convert_choice(correct_choice)\n" + doc_to_choice: "def doc_to_choice(doc):\n return [\" \" + convert_choice(doc[\"\ + choice1\"]), \" \" + convert_choice(doc[\"choice2\"])]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + indic_arc_challenge_hi: + task: indic_arc_challenge_hi + group: Cognitive-Lab/Indic-ARC-Challenge + dataset_path: Cognitive-Lab/Indic-ARC-Challenge + dataset_name: hi + test_split: test + doc_to_text: 'Question: {{translated_question}} + + Answer:' + doc_to_target: '{{translated_choices.label.index(answerKey)}}' + doc_to_choice: '{{translated_choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{translated_question}} + + Answer:' + metadata: + version: 1.0 + indic_arc_easy_hi: + task: indic_arc_easy_hi + group: Cognitive-Lab/Indic-ARC-Easy + dataset_path: Cognitive-Lab/Indic-ARC-Easy + dataset_name: hi + test_split: test + doc_to_text: 'Question: {{translated_question}} + + Answer:' + doc_to_target: '{{translated_choices.label.index(answerKey)}}' + doc_to_choice: '{{translated_choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{translated_question}} + + Answer:' + metadata: + version: 1.0 + indic_boolq_hi: + task: indic_boolq_hi + group: Cognitive-Lab/Indic-BoolQ + dataset_path: Cognitive-Lab/Indic-BoolQ + dataset_name: hi + validation_split: validation + doc_to_text: 'Passage: {translated_passage} + + Question: {translated_question.strip()} + + Answer:' + doc_to_target: answer + doc_to_choice: + - 'true' + - 'false' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + mrpc: + task: mrpc + group: glue + dataset_path: glue + dataset_name: mrpc + training_split: train + validation_split: validation + doc_to_text: 'Sentence 1: {{sentence1}} + + Sentence 2: {{sentence2}} + + Question: Do both sentences mean the same thing? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + - metric: f1 + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + piqa: + task: piqa + dataset_path: piqa + training_split: train + validation_split: validation + doc_to_text: 'Question: {{goal}} + + Answer:' + doc_to_target: label + doc_to_choice: '{{[sol1, sol2]}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: goal + metadata: + version: 1.0 + sst2: + task: sst2 + group: glue + dataset_path: glue + dataset_name: sst2 + training_split: train + validation_split: validation + doc_to_text: '{{sentence}} + + Question: Is this sentence positive or negative? + + Answer:' + doc_to_target: label + doc_to_choice: + - negative + - positive + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + winogrande: + task: winogrande + dataset_path: winogrande + dataset_name: winogrande_xl + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n answer_to_num = {\"1\": 0, \"2\": 1}\n\ + \ return answer_to_num[doc[\"answer\"]]\n" + doc_to_target: "def doc_to_target(doc):\n idx = doc[\"sentence\"].index(\"\ + _\") + 1\n return doc[\"sentence\"][idx:].strip()\n" + doc_to_choice: "def doc_to_choice(doc):\n idx = doc[\"sentence\"].index(\"\ + _\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"\ + sentence\"][:idx] + opt for opt in options]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: sentence + metadata: + version: 1.0 +cli_configs: + desc: null + value: + model: hf + model_args: pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step50000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer + batch_size: auto + batch_sizes: + - 64 + device: null + use_cache: null + limit: null + bootstrap_iters: 100000 + gen_kwargs: null diff --git a/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/media/table/evaluation/eval_results_1_5fe6a0b9b8eb34e9476c.table.json b/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/media/table/evaluation/eval_results_1_5fe6a0b9b8eb34e9476c.table.json new file mode 100644 index 0000000000000000000000000000000000000000..9df59cf114193b3ce33cb4e1109e55d930f6e98b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/media/table/evaluation/eval_results_1_5fe6a0b9b8eb34e9476c.table.json @@ -0,0 +1 @@ +{"columns": ["Tasks", "Version", "Filter", "num_fewshot", "Metric", "Value", "Stderr"], "data": [["winogrande", 1.0, "none", 0, "acc", "0.5122336227308603", "0.0140"], ["sst2", 1.0, "none", 0, "acc", "0.5022935779816514", "0.0169"], ["piqa", 1.0, "none", 0, "acc", "0.5043525571273123", "0.0117"], ["piqa", 1.0, "none", 0, "acc_norm", "0.500544069640914", "0.0117"], ["mrpc", 1.0, "none", 0, "acc", "0.3161764705882353", "0.0230"], ["mrpc", 1.0, "none", 0, "f1", "0.0", "0.0000"], ["indic_boolq_hi", 1.0, "none", 0, "acc", "0.6217125382262997", "0.0085"], ["indic_arc_easy_hi", 1.0, "none", 0, "acc", "0.25462962962962965", "0.0089"], ["indic_arc_challenge_hi", 1.0, "none", 0, "acc", "0.2090443686006826", "0.0119"], ["copa", 1.0, "none", 0, "acc", "0.59", "0.0494"], ["boolq", 2.0, "none", 0, "acc", "0.3782874617737003", "0.0085"], ["arc_easy", 1.0, "none", 0, "acc", "0.257996632996633", "0.0090"], ["arc_easy", 1.0, "none", 0, "acc_norm", "0.2638888888888889", "0.0090"]]} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/output.log b/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..d94c1315015b698876926783d012edec1f275e70 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240606_045127-vb760voe/files/output.log @@ -0,0 +1,730 @@ + +2024-06-06:04:51:27,826 INFO [__main__.py:251] Verbosity set to INFO +2024-06-06:04:51:37,038 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'boolq', 'copa', 'indic_arc_challenge_hi', 'indic_arc_easy_hi', 'indic_boolq_hi', 'mrpc', 'piqa', 'sst2', 'winogrande'] +2024-06-06:04:51:37,040 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-06-06:04:51:37,040 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step50000', 'tokenizer': '/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer'} +2024-06-06:04:51:39,516 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. +2024-06-06:04:52:05,950 WARNING [task.py:763] [Task: boolq] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-06:04:52:05,950 WARNING [task.py:775] [Task: boolq] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for super_glue contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/super_glue +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-06-06:04:52:08,694 WARNING [task.py:763] [Task: copa] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-06:04:52:08,695 WARNING [task.py:775] [Task: copa] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +2024-06-06:04:52:13,300 WARNING [task.py:322] [Task: indic_arc_challenge_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-06:04:52:13,301 WARNING [task.py:322] [Task: indic_arc_challenge_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-06:04:52:15,512 WARNING [task.py:322] [Task: indic_arc_easy_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-06:04:52:15,513 WARNING [task.py:322] [Task: indic_arc_easy_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-06:04:52:18,024 WARNING [task.py:763] [Task: mrpc] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-06:04:52:18,024 WARNING [task.py:775] [Task: mrpc] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +2024-06-06:04:52:18,025 WARNING [task.py:763] [Task: mrpc] metric f1 is defined, but aggregation is not. using default aggregation=f1 +2024-06-06:04:52:18,025 WARNING [task.py:775] [Task: mrpc] metric f1 is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for piqa contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/piqa +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-06-06:04:52:26,664 WARNING [task.py:763] [Task: sst2] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-06:04:52:26,665 WARNING [task.py:775] [Task: sst2] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for winogrande contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/winogrande +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-06-06:04:52:35,573 INFO [task.py:395] Building contexts for winogrande on rank 0... +100%|██████████| 1267/1267 [00:00<00:00, 69779.31it/s] +2024-06-06:04:52:35,660 INFO [task.py:395] Building contexts for sst2 on rank 0... +100%|██████████| 872/872 [00:00<00:00, 2589.63it/s] +2024-06-06:04:52:36,025 INFO [task.py:395] Building contexts for piqa on rank 0... +100%|██████████| 1838/1838 [00:01<00:00, 1103.01it/s] +2024-06-06:04:52:37,765 INFO [task.py:395] Building contexts for mrpc on rank 0... +100%|██████████| 408/408 [00:00<00:00, 1855.28it/s] +2024-06-06:04:52:38,004 INFO [task.py:395] Building contexts for indic_boolq_hi on rank 0... +100%|██████████| 3270/3270 [00:01<00:00, 3190.10it/s] +2024-06-06:04:52:39,202 INFO [task.py:395] Building contexts for indic_arc_easy_hi on rank 0... +100%|██████████| 2376/2376 [00:02<00:00, 1137.97it/s] +2024-06-06:04:52:41,524 INFO [task.py:395] Building contexts for indic_arc_challenge_hi on rank 0... +100%|██████████| 1172/1172 [00:01<00:00, 1139.83it/s] +2024-06-06:04:52:42,671 INFO [task.py:395] Building contexts for copa on rank 0... +100%|██████████| 100/100 [00:00<00:00, 61699.09it/s] +2024-06-06:04:52:42,680 INFO [task.py:395] Building contexts for boolq on rank 0... +100%|██████████| 3270/3270 [00:01<00:00, 2008.77it/s] +2024-06-06:04:52:44,443 INFO [task.py:395] Building contexts for arc_easy on rank 0... +100%|██████████| 2376/2376 [00:02<00:00, 1064.59it/s] +2024-06-06:04:52:46,819 INFO [evaluator.py:379] Running loglikelihood requests +Token indices sequence length is longer than the specified maximum sequence length for this model (1333 > 1024). Running this sequence through the model will result in indexing errors +Running loglikelihood requests: 0%| | 0/45739 [00:00