diff --git a/.gitattributes b/.gitattributes index 012f17aecd7eaf5a552523ae2a7ce8cfa2342109..2783cc1b32a0c3991289690c6431a6fc08b9dcf0 100644 --- a/.gitattributes +++ b/.gitattributes @@ -108,3 +108,4 @@ venv/lib/python3.10/site-packages/nvidia/cusolver/lib/libcusolver.so.11 filter=l venv/lib/python3.10/site-packages/pandas/_libs/tslibs/offsets.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text venv/lib/python3.10/site-packages/torch/lib/libtorch_cpu.so filter=lfs diff=lfs merge=lfs -text venv/bin/python filter=lfs diff=lfs merge=lfs -text +venv/lib/python3.10/site-packages/nvidia/cusparse/lib/libcusparse.so.12 filter=lfs diff=lfs merge=lfs -text diff --git a/ckpts/universal/global_step20/zero/17.attention.dense.weight/fp32.pt b/ckpts/universal/global_step20/zero/17.attention.dense.weight/fp32.pt new file mode 100644 index 0000000000000000000000000000000000000000..a53bebffeec53d59cbd8a0f1e5db386ef1e3cc91 --- /dev/null +++ b/ckpts/universal/global_step20/zero/17.attention.dense.weight/fp32.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc9c4fabd8ab1009d5a3060670179da1f515f243ce91449bc30e4fe0df8c083a +size 16778317 diff --git a/lm-evaluation-harness/tests/testdata/textsynth_test_434076260b6af3a46b7a5eaceec3306a5872c400a3872f744280b237455a0f8e.pkl b/lm-evaluation-harness/tests/testdata/textsynth_test_434076260b6af3a46b7a5eaceec3306a5872c400a3872f744280b237455a0f8e.pkl new file mode 100644 index 0000000000000000000000000000000000000000..09ac12f9334d8d77e64731eb53a258bc826e826b --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/textsynth_test_434076260b6af3a46b7a5eaceec3306a5872c400a3872f744280b237455a0f8e.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5b9ac677652592e9139c2623ad240c101dd337f7276de84c84062257c4d9b9a +size 2866 diff --git a/lm-evaluation-harness/tests/testdata/textsynth_test_4fd8d66a6dad7f602b40e5d7dc298d6fe329299d086a4659743a41f4a4012659.pkl b/lm-evaluation-harness/tests/testdata/textsynth_test_4fd8d66a6dad7f602b40e5d7dc298d6fe329299d086a4659743a41f4a4012659.pkl new file mode 100644 index 0000000000000000000000000000000000000000..39eba3fbdf7865fba5e076bc6c6b85418ef0c04a --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/textsynth_test_4fd8d66a6dad7f602b40e5d7dc298d6fe329299d086a4659743a41f4a4012659.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc36e5e5b1b97b2c588905c100bb7cbb5c262d1ed991ae182fe25cf8a88b61dd +size 1960 diff --git a/lm-evaluation-harness/tests/testdata/textsynth_test_7afdc285388e51094e12645f305328c759574fa3ec9751631025f8ad5ebf9f3e.pkl b/lm-evaluation-harness/tests/testdata/textsynth_test_7afdc285388e51094e12645f305328c759574fa3ec9751631025f8ad5ebf9f3e.pkl new file mode 100644 index 0000000000000000000000000000000000000000..94fc64bf12baf1c58b928641626e54d72d45b8f8 --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/textsynth_test_7afdc285388e51094e12645f305328c759574fa3ec9751631025f8ad5ebf9f3e.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a5612e9590402d33652e1664d1b5fa83cfb0f1fcaa25c782e314e8ddbcce231 +size 1975 diff --git a/lm-evaluation-harness/tests/testdata/textsynth_test_e7ad1e9f52a39e1ddd1e50f3c57ffa4546728dd150a67c0a0ddc8675c04e15d1.pkl b/lm-evaluation-harness/tests/testdata/textsynth_test_e7ad1e9f52a39e1ddd1e50f3c57ffa4546728dd150a67c0a0ddc8675c04e15d1.pkl new file mode 100644 index 0000000000000000000000000000000000000000..ae564e39b35d5553aac83af33c0dd126e3acf53e --- /dev/null +++ b/lm-evaluation-harness/tests/testdata/textsynth_test_e7ad1e9f52a39e1ddd1e50f3c57ffa4546728dd150a67c0a0ddc8675c04e15d1.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a052215e2933be5a3d5eb709fbf0d6136c25bd1bfef9ff0448165445bbce2afd +size 1848 diff --git a/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3fba2107e20f91e50460e6f6f06ce47b9e93aeb8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715682654 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/output.log b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e03f18ce4560e0abd29908ebe796c408235d7d66 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/output.log @@ -0,0 +1,42 @@ + +2024-05-14:10:30:54,713 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:10:30:59,157 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:10:30:59,159 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:10:30:59,159 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step120'} +Traceback (most recent call last): + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 398, in cached_file + resolved_file = hf_hub_download( + File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn + validate_repo_id(arg_value) + File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id + raise HFValidationError( +huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/data/cronscript/ckpts//hf_ckpt//global_step120'. Use `repo_type` argument if needed. +The above exception was the direct cause of the following exception: +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/data/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/data/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/data/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 928, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 631, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 686, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 462, in cached_file + raise EnvironmentError( +OSError: Incorrect path_or_model_id: '/data/cronscript/ckpts//hf_ckpt//global_step120'. Please provide either the path to a local folder or the repo_id of a model on the Hub. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..641623f3dfbe23cbc1b2136b43dc16f3aaf2c7dd --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T10:30:54.578276", + "startedAt": "2024-05-14T10:30:54.160257", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3392.447302631579, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.14, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.135, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3302.264, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.13, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3296.786, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 76.92245483398438 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 5}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8473613a5e33279837b9f13c7783f2eafd6d8230 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/logs/debug-internal.log @@ -0,0 +1,181 @@ +2024-05-14 10:30:54,174 INFO StreamThr :8254 [internal.py:wandb_internal():85] W&B internal server running at pid: 8254, started at: 2024-05-14 10:30:54.173429 +2024-05-14 10:30:54,175 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: status +2024-05-14 10:30:54,176 INFO WriterThread:8254 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/run-vz3w7dby.wandb +2024-05-14 10:30:54,177 DEBUG SenderThread:8254 [sender.py:send():378] send: header +2024-05-14 10:30:54,186 DEBUG SenderThread:8254 [sender.py:send():378] send: run +2024-05-14 10:30:54,442 INFO SenderThread:8254 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files +2024-05-14 10:30:54,442 INFO SenderThread:8254 [sender.py:_start_run_threads():1123] run started: vz3w7dby with start time 1715682654.172979 +2024-05-14 10:30:54,448 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 10:30:54,448 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: check_version +2024-05-14 10:30:54,529 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 10:30:54,530 DEBUG HandlerThread:8254 [system_info.py:__init__():26] System info init +2024-05-14 10:30:54,530 DEBUG HandlerThread:8254 [system_info.py:__init__():41] System info init done +2024-05-14 10:30:54,530 INFO HandlerThread:8254 [system_monitor.py:start():194] Starting system monitor +2024-05-14 10:30:54,530 INFO SystemMonitor:8254 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 10:30:54,530 INFO HandlerThread:8254 [system_monitor.py:probe():214] Collecting system info +2024-05-14 10:30:54,531 INFO SystemMonitor:8254 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 10:30:54,531 INFO SystemMonitor:8254 [interfaces.py:start():188] Started disk monitoring +2024-05-14 10:30:54,531 INFO SystemMonitor:8254 [interfaces.py:start():188] Started memory monitoring +2024-05-14 10:30:54,532 INFO SystemMonitor:8254 [interfaces.py:start():188] Started network monitoring +2024-05-14 10:30:54,578 DEBUG HandlerThread:8254 [system_info.py:probe():150] Probing system +2024-05-14 10:30:54,587 DEBUG HandlerThread:8254 [system_info.py:_probe_git():135] Probing git +2024-05-14 10:30:54,607 ERROR HandlerThread:8254 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 10:30:54,607 DEBUG HandlerThread:8254 [system_info.py:_probe_git():143] Probing git done +2024-05-14 10:30:54,607 DEBUG HandlerThread:8254 [system_info.py:probe():198] Probing system done +2024-05-14 10:30:54,607 DEBUG HandlerThread:8254 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T10:30:54.578276', 'startedAt': '2024-05-14T10:30:54.160257', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3392.447302631579, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.14, 'min': 800.0, 'max': 3400.0}, {'current': 3300.135, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3302.264, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.13, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3296.786, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 76.92245483398438}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 10:30:54,607 INFO HandlerThread:8254 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 10:30:54,607 INFO HandlerThread:8254 [system_monitor.py:probe():227] Publishing system info +2024-05-14 10:30:54,608 INFO HandlerThread:8254 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 10:30:54,612 DEBUG SenderThread:8254 [sender.py:send():378] send: files +2024-05-14 10:30:54,612 INFO SenderThread:8254 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 10:30:54,709 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 10:30:54,709 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: python_packages +2024-05-14 10:30:54,709 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 10:30:54,710 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: stop_status +2024-05-14 10:30:54,875 DEBUG SenderThread:8254 [sender.py:send():378] send: telemetry +2024-05-14 10:30:55,126 INFO wandb-upload_0:8254 [upload_job.py:push():130] Uploaded file /tmp/tmpdqnvjmg9wandb/8e8d6l51-wandb-metadata.json +2024-05-14 10:30:55,443 INFO Thread-12 :8254 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-metadata.json +2024-05-14 10:30:55,444 INFO Thread-12 :8254 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/requirements.txt +2024-05-14 10:30:55,444 INFO Thread-12 :8254 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/output.log +2024-05-14 10:30:57,446 INFO Thread-12 :8254 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/output.log +2024-05-14 10:31:00,159 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:31:00,224 DEBUG SenderThread:8254 [sender.py:send():378] send: exit +2024-05-14 10:31:00,224 INFO SenderThread:8254 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 10:31:00,224 INFO SenderThread:8254 [sender.py:send_exit():587] handling runtime: 5 +2024-05-14 10:31:00,225 INFO SenderThread:8254 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:31:00,226 INFO SenderThread:8254 [sender.py:send_exit():593] send defer +2024-05-14 10:31:00,226 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,226 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 10:31:00,226 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,226 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 10:31:00,226 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 1 +2024-05-14 10:31:00,226 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,226 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 10:31:00,226 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,226 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 10:31:00,226 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 2 +2024-05-14 10:31:00,226 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,226 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 10:31:00,226 INFO HandlerThread:8254 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 10:31:00,226 DEBUG SystemMonitor:8254 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 10:31:00,227 INFO HandlerThread:8254 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 10:31:00,227 DEBUG SystemMonitor:8254 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 10:31:00,227 INFO HandlerThread:8254 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 10:31:00,227 DEBUG SystemMonitor:8254 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 10:31:00,227 INFO HandlerThread:8254 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 10:31:00,228 INFO HandlerThread:8254 [interfaces.py:finish():200] Joined network monitor +2024-05-14 10:31:00,229 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,229 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 10:31:00,229 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 3 +2024-05-14 10:31:00,229 DEBUG SenderThread:8254 [sender.py:send():378] send: stats +2024-05-14 10:31:00,229 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,229 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 10:31:00,229 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,229 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 10:31:00,230 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 4 +2024-05-14 10:31:00,230 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,230 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 10:31:00,230 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,230 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 10:31:00,230 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 5 +2024-05-14 10:31:00,230 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,230 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 10:31:00,230 DEBUG SenderThread:8254 [sender.py:send():378] send: summary +2024-05-14 10:31:00,231 INFO SenderThread:8254 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:31:00,231 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,231 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 10:31:00,231 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 6 +2024-05-14 10:31:00,231 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,231 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 10:31:00,231 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,231 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 10:31:00,233 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:31:00,306 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 7 +2024-05-14 10:31:00,306 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,306 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 10:31:00,306 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,306 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 10:31:00,448 INFO Thread-12 :8254 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/config.yaml +2024-05-14 10:31:00,448 INFO Thread-12 :8254 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-summary.json +2024-05-14 10:31:00,888 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 8 +2024-05-14 10:31:00,888 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,888 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 10:31:00,888 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,888 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 10:31:00,888 INFO SenderThread:8254 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 10:31:00,889 INFO SenderThread:8254 [job_builder.py:_get_source_type():576] no source found +2024-05-14 10:31:00,889 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 9 +2024-05-14 10:31:00,889 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,889 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 10:31:00,889 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,889 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 10:31:00,889 INFO SenderThread:8254 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 10:31:01,224 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:01,449 INFO SenderThread:8254 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/output.log +2024-05-14 10:31:01,449 INFO SenderThread:8254 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files +2024-05-14 10:31:01,449 INFO SenderThread:8254 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-summary.json wandb-summary.json +2024-05-14 10:31:01,450 INFO SenderThread:8254 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/config.yaml config.yaml +2024-05-14 10:31:01,450 INFO SenderThread:8254 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/requirements.txt requirements.txt +2024-05-14 10:31:01,450 INFO SenderThread:8254 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/output.log output.log +2024-05-14 10:31:01,450 INFO SenderThread:8254 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-metadata.json wandb-metadata.json +2024-05-14 10:31:01,450 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 10 +2024-05-14 10:31:01,450 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:01,450 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:01,450 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 10:31:01,451 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:01,452 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 10:31:01,452 INFO SenderThread:8254 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:31:01,684 INFO wandb-upload_0:8254 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/wandb-summary.json +2024-05-14 10:31:01,863 INFO wandb-upload_2:8254 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/requirements.txt +2024-05-14 10:31:01,944 INFO wandb-upload_3:8254 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/output.log +2024-05-14 10:31:01,954 INFO wandb-upload_1:8254 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/files/config.yaml +2024-05-14 10:31:02,155 INFO Thread-11 (_thread_body):8254 [sender.py:transition_state():613] send defer: 11 +2024-05-14 10:31:02,155 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:02,155 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 10:31:02,155 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:02,156 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 10:31:02,156 INFO SenderThread:8254 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 10:31:02,156 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 12 +2024-05-14 10:31:02,156 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:02,156 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 10:31:02,156 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:02,156 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 10:31:02,156 INFO SenderThread:8254 [file_stream.py:finish():601] file stream finish called +2024-05-14 10:31:02,225 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:02,360 INFO SenderThread:8254 [file_stream.py:finish():605] file stream finish is done +2024-05-14 10:31:02,360 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 13 +2024-05-14 10:31:02,360 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:02,360 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:02,360 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 10:31:02,361 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:02,361 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 10:31:02,361 INFO SenderThread:8254 [sender.py:transition_state():613] send defer: 14 +2024-05-14 10:31:02,361 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:02,361 DEBUG SenderThread:8254 [sender.py:send():378] send: final +2024-05-14 10:31:02,361 INFO HandlerThread:8254 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 10:31:02,361 DEBUG SenderThread:8254 [sender.py:send():378] send: footer +2024-05-14 10:31:02,361 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:02,361 INFO SenderThread:8254 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 10:31:02,361 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:02,361 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:02,362 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:02,362 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:02,362 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 10:31:02,362 DEBUG SenderThread:8254 [sender.py:send_request():405] send_request: server_info +2024-05-14 10:31:02,363 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 10:31:02,363 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 10:31:02,364 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 10:31:02,415 INFO MainThread:8254 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 10:31:02,415 INFO MainThread:8254 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 10:31:02,415 INFO MainThread:8254 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 10:31:02,415 DEBUG HandlerThread:8254 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 10:31:02,415 INFO HandlerThread:8254 [handler.py:finish():882] shutting down handler +2024-05-14 10:31:03,362 INFO WriterThread:8254 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/run-vz3w7dby.wandb +2024-05-14 10:31:03,415 INFO SenderThread:8254 [sender.py:finish():1545] shutting down sender +2024-05-14 10:31:03,415 INFO SenderThread:8254 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:31:03,415 INFO SenderThread:8254 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..e30254010a474075aadec371a8713285d42d88ca --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_setup.py:_flush():76] Configure stats pid to 6974 +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 10:30:54,170 WARNING MainThread:6974 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/logs/debug.log +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/logs/debug-internal.log +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_init.py:init():560] calling init triggers +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_init.py:init():610] starting backend +2024-05-14 10:30:54,170 INFO MainThread:6974 [wandb_init.py:init():614] setting up manager +2024-05-14 10:30:54,172 INFO MainThread:6974 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 10:30:54,172 INFO MainThread:6974 [wandb_init.py:init():622] backend started and connected +2024-05-14 10:30:54,175 INFO MainThread:6974 [wandb_init.py:init():711] updated telemetry +2024-05-14 10:30:54,186 INFO MainThread:6974 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 10:30:54,448 INFO MainThread:6974 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 10:30:54,524 INFO MainThread:6974 [wandb_run.py:_on_init():2405] got version response +2024-05-14 10:30:54,525 INFO MainThread:6974 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 10:30:54,710 INFO MainThread:6974 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 10:30:54,710 INFO MainThread:6974 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 10:30:54,710 INFO MainThread:6974 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 10:30:54,710 INFO MainThread:6974 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 10:30:54,712 INFO MainThread:6974 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 10:31:03,416 WARNING MsgRouterThr:6974 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/run-vz3w7dby.wandb b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/run-vz3w7dby.wandb new file mode 100644 index 0000000000000000000000000000000000000000..91111918a4f076e36c876e1f162066a786286138 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_103054-vz3w7dby/run-vz3w7dby.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..299eb55ae9fb758d9b2420d75d18ee2816d408fb --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/files/config.yaml @@ -0,0 +1,32 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715686992 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..ab5463c815e428c4cd7097e10105c78e20d65d6b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/logs/debug-internal.log @@ -0,0 +1,19 @@ +2024-05-14 11:43:12,808 INFO StreamThr :80505 [internal.py:wandb_internal():85] W&B internal server running at pid: 80505, started at: 2024-05-14 11:43:12.807900 +2024-05-14 11:43:12,810 DEBUG HandlerThread:80505 [handler.py:handle_request():158] handle_request: status +2024-05-14 11:43:12,811 INFO WriterThread:80505 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/run-fv5jw0zy.wandb +2024-05-14 11:43:12,812 DEBUG SenderThread:80505 [sender.py:send():378] send: header +2024-05-14 11:43:12,821 DEBUG SenderThread:80505 [sender.py:send():378] send: run +2024-05-14 11:43:13,062 INFO SenderThread:80505 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/files +2024-05-14 11:43:13,062 INFO SenderThread:80505 [sender.py:_start_run_threads():1123] run started: fv5jw0zy with start time 1715686992.807246 +2024-05-14 11:43:13,811 WARNING StreamThr :80505 [internal.py:is_dead():413] Internal process exiting, parent pid 79482 disappeared +2024-05-14 11:43:13,811 ERROR StreamThr :80505 [internal.py:wandb_internal():151] Internal process shutdown. +2024-05-14 11:43:13,821 INFO HandlerThread:80505 [handler.py:finish():882] shutting down handler +2024-05-14 11:43:13,821 INFO WriterThread:80505 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/run-fv5jw0zy.wandb +2024-05-14 11:43:14,062 INFO SenderThread:80505 [sender.py:finish():1545] shutting down sender +2024-05-14 11:43:14,062 INFO SenderThread:80505 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 11:43:15,063 INFO SenderThread:80505 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/files +2024-05-14 11:43:15,063 INFO SenderThread:80505 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/files/config.yaml config.yaml +2024-05-14 11:43:15,063 INFO SenderThread:80505 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 11:43:15,063 INFO SenderThread:80505 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 11:43:15,535 INFO wandb-upload_0:80505 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/files/config.yaml +2024-05-14 11:43:15,811 INFO MainThread:80505 [internal.py:handle_exit():75] Internal process exited diff --git a/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..5a6e3af8391d3d0740ffef6750b27d8a03f2d30f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/logs/debug.log @@ -0,0 +1,20 @@ +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_setup.py:_flush():76] Configure stats pid to 79482 +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 11:43:12,804 WARNING MainThread:79482 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/logs/debug.log +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/logs/debug-internal.log +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_init.py:init():560] calling init triggers +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 11:43:12,804 INFO MainThread:79482 [wandb_init.py:init():610] starting backend +2024-05-14 11:43:12,805 INFO MainThread:79482 [wandb_init.py:init():614] setting up manager +2024-05-14 11:43:12,806 INFO MainThread:79482 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 11:43:12,807 INFO MainThread:79482 [wandb_init.py:init():622] backend started and connected +2024-05-14 11:43:12,809 INFO MainThread:79482 [wandb_init.py:init():711] updated telemetry +2024-05-14 11:43:12,820 INFO MainThread:79482 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout diff --git a/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/run-fv5jw0zy.wandb b/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/run-fv5jw0zy.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ef6b59fe6bfe6fd032d32277c92c479a113ed342 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_114312-fv5jw0zy/run-fv5jw0zy.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..47c0abcf13348614252629c526b475cedbfa8165 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/config.yaml @@ -0,0 +1,86 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715704468 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 2 + - 23 + - 62 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 +task_configs: + desc: null + value: + indiccopa-hi: + task: indiccopa-hi + group: ai4bharat/IndicCOPA + dataset_path: ai4bharat/IndicCOPA + dataset_name: translation-hi + test_split: test + doc_to_text: 'functools.partial(, connector={''cause'': + ''कारण'', ''effect'': ''परिणाम''})' + doc_to_target: label + doc_to_choice: "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"\ + ]), convert_choice(doc[\"choice2\"])]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 +cli_configs: + desc: null + value: + model: hf + model_args: pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100 + batch_size: auto + batch_sizes: + - 64 + device: null + use_cache: null + limit: null + bootstrap_iters: 100000 + gen_kwargs: null diff --git a/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json b/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json new file mode 100644 index 0000000000000000000000000000000000000000..fc9a92c5b80c0631baf44f28a9d55b68d269f5da --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json @@ -0,0 +1 @@ +{"columns": ["Tasks", "Version", "Filter", "num_fewshot", "Metric", "Value", "Stderr"], "data": [["indiccopa-hi", 1.0, "none", 0, "acc", "0.534521158129176", "0.0236"]]} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/output.log b/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..68fa60f5616a76c3571201be194569b419980e7f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/files/output.log @@ -0,0 +1,33 @@ + +2024-05-14:16:34:29,263 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:16:34:34,001 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:16:34:34,003 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:16:34:34,003 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step100'} +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +[2024-05-14 16:34:40,705] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +2024-05-14:16:34:41,102 INFO [huggingface.py:334] Using 8 devices with data parallelism +2024-05-14:16:34:41,106 WARNING [task.py:763] [Task: indiccopa-hi] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-14:16:34:41,106 WARNING [task.py:775] [Task: indiccopa-hi] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for ai4bharat/IndicCOPA contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/ai4bharat/IndicCOPA +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-05-14:16:34:42,518 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:16:34:42,518 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:16:34:42,538 INFO [task.py:395] Building contexts for indiccopa-hi on rank 0... +100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 57/57 [00:00<00:00, 98914.08it/s] +2024-05-14:16:34:44,718 INFO [evaluator.py:379] Running loglikelihood requests +Running loglikelihood requests: 0%| | 0/114 [00:00, connector={'cause': 'कारण', 'effect': 'परिणाम'})", 'doc_to_target': 'label', 'doc_to_choice': 'def doc_to_choice(doc):\n return [convert_choice(doc["choice1"]), convert_choice(doc["choice2"])]\n', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}}, 'cli_configs': {'model': 'hf', 'model_args': 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100', 'batch_size': 'auto', 'batch_sizes': [64], 'device': None, 'use_cache': None, 'limit': None, 'bootstrap_iters': 100000, 'gen_kwargs': None}} +2024-05-14 16:34:53,684 INFO MainThread:116792 [wandb_run.py:_finish():2103] finishing run smlgenai/bharatgpt/9ixw6vnd +2024-05-14 16:34:53,684 INFO MainThread:116792 [wandb_run.py:_atexit_cleanup():2343] got exitcode: 0 +2024-05-14 16:34:53,685 INFO MainThread:116792 [wandb_run.py:_restore():2326] restore +2024-05-14 16:34:53,685 INFO MainThread:116792 [wandb_run.py:_restore():2332] restore done +2024-05-14 16:35:00,927 INFO MainThread:116792 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 16:35:00,927 INFO MainThread:116792 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 16:35:00,931 INFO MainThread:116792 [wandb_run.py:_footer_sync_info():3953] logging synced files diff --git a/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/run-9ixw6vnd.wandb b/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/run-9ixw6vnd.wandb new file mode 100644 index 0000000000000000000000000000000000000000..8837fc8ec8966bae5e9aa6759b77f046e02c1485 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_163428-9ixw6vnd/run-9ixw6vnd.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..77d2fc6b1aadfaa06d48159f85f9b4ec7d748e71 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/config.yaml @@ -0,0 +1,86 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716384490 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 2 + - 23 + - 62 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 +task_configs: + desc: null + value: + indiccopa-hi: + task: indiccopa-hi + group: ai4bharat/IndicCOPA + dataset_path: ai4bharat/IndicCOPA + dataset_name: translation-hi + test_split: test + doc_to_text: 'functools.partial(, connector={''cause'': + ''कारण'', ''effect'': ''परिणाम''})' + doc_to_target: label + doc_to_choice: "def doc_to_choice(doc):\n return [convert_choice(doc[\"choice1\"\ + ]), convert_choice(doc[\"choice2\"])]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 +cli_configs: + desc: null + value: + model: hf + model_args: pretrained=/mnt/weka/peacock/idc/cronscript/ckpts//hf_ckpt//global_step100 + batch_size: auto + batch_sizes: + - 64 + device: null + use_cache: null + limit: null + bootstrap_iters: 100000 + gen_kwargs: null diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json new file mode 100644 index 0000000000000000000000000000000000000000..fc9a92c5b80c0631baf44f28a9d55b68d269f5da --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json @@ -0,0 +1 @@ +{"columns": ["Tasks", "Version", "Filter", "num_fewshot", "Metric", "Value", "Stderr"], "data": [["indiccopa-hi", 1.0, "none", 0, "acc", "0.534521158129176", "0.0236"]]} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..10f30c300626c45dca8b0d2cef63de2ba5b6615e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log @@ -0,0 +1,51 @@ + +2024-05-22:13:28:11,357 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:13:28:20,939 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-22:13:28:20,940 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:13:28:20,941 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/idc/cronscript/ckpts//hf_ckpt//global_step100'} +2024-05-22:13:28:23,273 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +2024-05-22:13:28:31,525 WARNING [task.py:763] [Task: indiccopa-hi] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-22:13:28:31,525 WARNING [task.py:775] [Task: indiccopa-hi] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for ai4bharat/IndicCOPA contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/ai4bharat/IndicCOPA +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 2.30k/2.30k [00:00<00:00, 5.60MB/s] +Downloading readme: 100%|██████████| 2.89k/2.89k [00:00<00:00, 7.52MB/s] +Downloading data: 33%|███▎ | 50.2k/152k [00:00<00:00, 397kB/s] +Passed argument batch_size = auto:1. Detecting largest batch size +Downloading data: 100%|██████████| 152k/152k [00:00<00:00, 823kB/s] +Generating test split: 449 examples [00:00, 12480.98 examples/s] +2024-05-22:13:28:35,121 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-22:13:28:35,121 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-22:13:28:35,153 INFO [task.py:395] Building contexts for indiccopa-hi on rank 0... +100%|██████████| 449/449 [00:00<00:00, 78321.58it/s] +2024-05-22:13:28:35,204 INFO [evaluator.py:379] Running loglikelihood requests + + + + + + + + + + + + + + + + + + + +Running loglikelihood requests: 100%|██████████| 898/898 [01:17<00:00, 11.56it/s] +hf (pretrained=/mnt/weka/peacock/idc/cronscript/ckpts//hf_ckpt//global_step100), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: auto (64) +| Tasks |Version|Filter|n-shot|Metric|Value | |Stderr| +|------------|------:|------|-----:|------|-----:|---|-----:| +|indiccopa-hi| 1|none | 0|acc |0.5345|± |0.0236| \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..69a03246f3370f77f9b9930e80c184d14e127c2b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T13:28:11.145756", + "startedAt": "2024-05-22T13:28:10.603025", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/idc/cronscript/ckpts//hf_ckpt//global_step100", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_1" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2340.46931875, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.003, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3347.346, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3341.534, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.63244247436523 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..307cf3a4546a0cbbe43a5331c73a0fb5dc57db96 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-summary.json @@ -0,0 +1 @@ +{"indiccopa-hi/alias": "indiccopa-hi", "indiccopa-hi/acc": 0.534521158129176, "indiccopa-hi/acc_stderr": 0.023566409283576152, "_timestamp": 1716384594.8208838, "_runtime": 104.19854474067688, "_step": 1, "evaluation/eval_results": {"_type": "table-file", "sha256": "c78c57917f215d296d9d79a9c088f03a0c26963e80eec711a3cc1777734a1c37", "size": 169, "artifact_path": "wandb-client-artifact://jashb8tixn54tc7ml36kjkm743fxtdk4k70fmmcuv2p04qhedgxh4tn9j7uxw41k2taksyt4acg332ih30v35chrxb97tbt3puizqd0eigp7j5hfj5jjosqrxdkptqiv/evaluation/eval_results.table.json", "_latest_artifact_path": "wandb-client-artifact://2k0lh2sih6gwixpkvzfei4jryg321dok76e7hozt77q5i7grmx80brd8u6i1n1w14tjx7xty7hhkhyw9x0n4mpldhh7sow48qa2weo53xc42vi5gyksnqd9iba87apie:latest/evaluation/eval_results.table.json", "path": "media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json", "ncols": 7, "nrows": 1}, "_wandb": {"runtime": 104}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..1cd2b1db877c619c21e914db8e9b01ea8ce6190d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/logs/debug-internal.log @@ -0,0 +1,269 @@ +2024-05-22 13:28:10,624 INFO StreamThr :825 [internal.py:wandb_internal():85] W&B internal server running at pid: 825, started at: 2024-05-22 13:28:10.622167 +2024-05-22 13:28:10,628 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status +2024-05-22 13:28:10,631 INFO WriterThread:825 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/run-pdly1svl.wandb +2024-05-22 13:28:10,633 DEBUG SenderThread:825 [sender.py:send():378] send: header +2024-05-22 13:28:10,635 DEBUG SenderThread:825 [sender.py:send():378] send: run +2024-05-22 13:28:10,919 INFO SenderThread:825 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files +2024-05-22 13:28:10,920 INFO SenderThread:825 [sender.py:_start_run_threads():1123] run started: pdly1svl with start time 1716384490.622339 +2024-05-22 13:28:10,926 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 13:28:10,926 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: check_version +2024-05-22 13:28:11,048 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 13:28:11,050 DEBUG HandlerThread:825 [system_info.py:__init__():26] System info init +2024-05-22 13:28:11,050 DEBUG HandlerThread:825 [system_info.py:__init__():41] System info init done +2024-05-22 13:28:11,050 INFO HandlerThread:825 [system_monitor.py:start():194] Starting system monitor +2024-05-22 13:28:11,050 INFO SystemMonitor:825 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 13:28:11,050 INFO HandlerThread:825 [system_monitor.py:probe():214] Collecting system info +2024-05-22 13:28:11,057 INFO SystemMonitor:825 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 13:28:11,058 INFO SystemMonitor:825 [interfaces.py:start():188] Started disk monitoring +2024-05-22 13:28:11,062 INFO SystemMonitor:825 [interfaces.py:start():188] Started memory monitoring +2024-05-22 13:28:11,063 INFO SystemMonitor:825 [interfaces.py:start():188] Started network monitoring +2024-05-22 13:28:11,145 DEBUG HandlerThread:825 [system_info.py:probe():150] Probing system +2024-05-22 13:28:11,149 DEBUG HandlerThread:825 [system_info.py:_probe_git():135] Probing git +2024-05-22 13:28:11,159 ERROR HandlerThread:825 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 13:28:11,159 DEBUG HandlerThread:825 [system_info.py:_probe_git():143] Probing git done +2024-05-22 13:28:11,160 DEBUG HandlerThread:825 [system_info.py:probe():198] Probing system done +2024-05-22 13:28:11,160 DEBUG HandlerThread:825 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T13:28:11.145756', 'startedAt': '2024-05-22T13:28:10.603025', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/idc/cronscript/ckpts//hf_ckpt//global_step100', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_1'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2340.46931875, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.003, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3347.346, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3341.534, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.63244247436523}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 13:28:11,160 INFO HandlerThread:825 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 13:28:11,160 INFO HandlerThread:825 [system_monitor.py:probe():227] Publishing system info +2024-05-22 13:28:11,163 INFO HandlerThread:825 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 13:28:11,169 DEBUG SenderThread:825 [sender.py:send():378] send: files +2024-05-22 13:28:11,169 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 13:28:11,352 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 13:28:11,352 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: python_packages +2024-05-22 13:28:11,354 DEBUG SenderThread:825 [sender.py:send():378] send: telemetry +2024-05-22 13:28:11,355 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 13:28:11,355 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-22 13:28:11,782 INFO wandb-upload_0:825 [upload_job.py:push():130] Uploaded file /tmp/tmpsxxb7pxywandb/8z6zhuhg-wandb-metadata.json +2024-05-22 13:28:11,921 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:11,922 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/requirements.txt +2024-05-22 13:28:11,922 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-metadata.json +2024-05-22 13:28:13,921 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:16,501 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:28:21,926 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:21,941 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:28:25,931 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:26,353 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 13:28:26,354 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-22 13:28:27,489 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:28:27,967 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:32,528 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:28:34,054 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:36,123 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:37,784 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:28:38,444 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:41,353 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 13:28:41,353 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-22 13:28:42,957 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:28:44,258 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:44,259 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/config.yaml +2024-05-22 13:28:45,610 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:48,412 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:28:50,222 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:52,621 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:54,179 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:28:56,048 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:28:56,353 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 13:28:56,354 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-22 13:28:59,491 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:02,060 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:04,611 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:07,602 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:10,238 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:11,063 DEBUG SystemMonitor:825 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 13:29:11,066 DEBUG SenderThread:825 [sender.py:send():378] send: stats +2024-05-22 13:29:11,353 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 13:29:11,353 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-22 13:29:13,821 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:15,677 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:18,255 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:20,283 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:21,150 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:23,614 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:26,353 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 13:29:26,354 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-22 13:29:26,437 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:28,404 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:30,707 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:31,871 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:34,305 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:36,908 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:37,696 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:41,068 DEBUG SenderThread:825 [sender.py:send():378] send: stats +2024-05-22 13:29:41,353 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 13:29:41,354 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-22 13:29:42,537 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:43,816 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:48,216 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:48,707 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:52,132 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:54,067 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:54,134 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:54,403 DEBUG SenderThread:825 [sender.py:send():378] send: config +2024-05-22 13:29:54,404 DEBUG SenderThread:825 [sender.py:send():378] send: telemetry +2024-05-22 13:29:54,405 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: summary_record +2024-05-22 13:29:54,405 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: partial_history +2024-05-22 13:29:54,406 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 13:29:54,407 DEBUG SenderThread:825 [sender.py:send():378] send: history +2024-05-22 13:29:54,407 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: summary_record +2024-05-22 13:29:54,408 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 13:29:54,609 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: log_artifact +2024-05-22 13:29:54,610 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: log_artifact +2024-05-22 13:29:54,821 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: partial_history +2024-05-22 13:29:54,977 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: log_artifact +2024-05-22 13:29:55,137 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-summary.json +2024-05-22 13:29:55,138 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json +2024-05-22 13:29:55,138 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/media +2024-05-22 13:29:55,138 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/media/table +2024-05-22 13:29:55,138 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/media/table/evaluation +2024-05-22 13:29:55,377 INFO wandb-upload_0:825 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpjhha0nue +2024-05-22 13:29:55,640 INFO wandb-upload_1:825 [upload_job.py:push():130] Uploaded file /tmp/tmpsxxb7pxywandb/68m93s6z-media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json +2024-05-22 13:29:55,910 INFO SenderThread:825 [sender.py:send_request_log_artifact():1455] logged artifact run-pdly1svl-evaluationeval_results - {'id': 'QXJ0aWZhY3Q6ODQ2MTQ2ODQ3', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3ODAwNjA3Nw==', 'latestArtifact': None}} +2024-05-22 13:29:55,910 DEBUG SenderThread:825 [sender.py:send():378] send: files +2024-05-22 13:29:55,910 INFO SenderThread:825 [sender.py:_save_file():1389] saving file media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json with policy now +2024-05-22 13:29:55,911 DEBUG SenderThread:825 [sender.py:send():378] send: history +2024-05-22 13:29:55,911 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: summary_record +2024-05-22 13:29:55,914 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 13:29:55,914 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: log_artifact +2024-05-22 13:29:56,062 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 13:29:56,140 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-summary.json +2024-05-22 13:29:56,227 INFO wandb-upload_0:825 [upload_job.py:push():130] Uploaded file /tmp/tmpsxxb7pxywandb/w5oieqyy-media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json +2024-05-22 13:29:56,543 INFO wandb-upload_1:825 [upload_job.py:push():88] Uploaded file /tmp/tmp4gmrbk1s/results.json +2024-05-22 13:29:57,111 INFO SenderThread:825 [sender.py:send_request_log_artifact():1455] logged artifact results - {'id': 'QXJ0aWZhY3Q6ODQ2MTQ2ODY2', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE3NDY1ODEyNw==', 'latestArtifact': {'id': 'QXJ0aWZhY3Q6ODM2MjE5ODEy', 'versionIndex': 4}}} +2024-05-22 13:29:57,111 DEBUG SenderThread:825 [sender.py:send():378] send: telemetry +2024-05-22 13:29:57,111 DEBUG SenderThread:825 [sender.py:send():378] send: exit +2024-05-22 13:29:57,111 INFO SenderThread:825 [sender.py:send_exit():585] handling exit code: 0 +2024-05-22 13:29:57,112 INFO SenderThread:825 [sender.py:send_exit():587] handling runtime: 104 +2024-05-22 13:29:57,113 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 13:29:57,113 INFO SenderThread:825 [sender.py:send_exit():593] send defer +2024-05-22 13:29:57,113 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 13:29:57,113 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:57,113 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 13:29:57,113 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:57,113 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 13:29:57,114 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 1 +2024-05-22 13:29:57,114 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:57,114 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 13:29:57,114 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:57,114 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 13:29:57,114 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 2 +2024-05-22 13:29:57,114 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:57,114 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 13:29:57,114 INFO HandlerThread:825 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 13:29:57,114 DEBUG SystemMonitor:825 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 13:29:57,114 DEBUG SystemMonitor:825 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 13:29:57,117 INFO HandlerThread:825 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 13:29:57,117 INFO HandlerThread:825 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 13:29:57,117 INFO HandlerThread:825 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 13:29:57,117 INFO HandlerThread:825 [interfaces.py:finish():200] Joined network monitor +2024-05-22 13:29:57,118 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:57,118 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 13:29:57,118 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 3 +2024-05-22 13:29:57,118 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:57,118 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 13:29:57,118 DEBUG SenderThread:825 [sender.py:send():378] send: stats +2024-05-22 13:29:57,119 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:57,119 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 13:29:57,119 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 4 +2024-05-22 13:29:57,119 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:57,119 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 13:29:57,119 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:57,119 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 13:29:57,119 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 5 +2024-05-22 13:29:57,120 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:57,120 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 13:29:57,120 DEBUG SenderThread:825 [sender.py:send():378] send: summary +2024-05-22 13:29:57,121 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 13:29:57,121 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:57,121 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 13:29:57,121 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 6 +2024-05-22 13:29:57,121 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:57,121 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 13:29:57,121 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:57,121 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 13:29:57,125 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 13:29:57,141 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-summary.json +2024-05-22 13:29:57,225 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 7 +2024-05-22 13:29:57,225 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:57,225 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 13:29:57,225 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:57,225 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 13:29:58,063 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 13:29:58,146 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:29:58,146 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/config.yaml +2024-05-22 13:29:59,480 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 8 +2024-05-22 13:29:59,480 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 13:29:59,480 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:59,480 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 13:29:59,480 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:59,480 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 13:29:59,480 INFO SenderThread:825 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 13:29:59,481 INFO SenderThread:825 [job_builder.py:_get_source_type():576] no source found +2024-05-22 13:29:59,481 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 9 +2024-05-22 13:29:59,482 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:29:59,482 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 13:29:59,482 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:29:59,482 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 13:29:59,482 INFO SenderThread:825 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 13:30:00,063 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 13:30:00,148 INFO SenderThread:825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:30:00,149 INFO SenderThread:825 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files +2024-05-22 13:30:00,149 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log output.log +2024-05-22 13:30:00,150 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-summary.json wandb-summary.json +2024-05-22 13:30:00,152 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-metadata.json wandb-metadata.json +2024-05-22 13:30:00,154 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/config.yaml config.yaml +2024-05-22 13:30:00,154 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/requirements.txt requirements.txt +2024-05-22 13:30:00,155 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json media/table/evaluation/eval_results_1_c78c57917f215d296d9d.table.json +2024-05-22 13:30:00,155 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 10 +2024-05-22 13:30:00,155 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 13:30:00,155 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:30:00,156 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 13:30:00,156 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:30:00,156 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 13:30:00,156 INFO SenderThread:825 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 13:30:00,394 INFO wandb-upload_0:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/output.log +2024-05-22 13:30:00,399 INFO wandb-upload_1:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/wandb-summary.json +2024-05-22 13:30:00,688 INFO wandb-upload_2:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/config.yaml +2024-05-22 13:30:00,751 INFO wandb-upload_3:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/files/requirements.txt +2024-05-22 13:30:00,951 INFO Thread-11 (_thread_body):825 [sender.py:transition_state():613] send defer: 11 +2024-05-22 13:30:00,951 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:30:00,951 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 13:30:00,951 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:30:00,951 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 13:30:00,951 INFO SenderThread:825 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 13:30:00,952 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 12 +2024-05-22 13:30:00,952 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:30:00,952 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 13:30:00,952 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:30:00,952 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 13:30:00,952 INFO SenderThread:825 [file_stream.py:finish():601] file stream finish called +2024-05-22 13:30:01,031 INFO SenderThread:825 [file_stream.py:finish():605] file stream finish is done +2024-05-22 13:30:01,031 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 13 +2024-05-22 13:30:01,032 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:30:01,032 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 13:30:01,032 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:30:01,032 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 13:30:01,032 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 14 +2024-05-22 13:30:01,032 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-22 13:30:01,032 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 13:30:01,032 DEBUG SenderThread:825 [sender.py:send():378] send: final +2024-05-22 13:30:01,032 DEBUG SenderThread:825 [sender.py:send():378] send: footer +2024-05-22 13:30:01,032 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-22 13:30:01,032 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 13:30:01,033 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 13:30:01,033 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 13:30:01,034 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 13:30:01,034 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 13:30:01,034 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: server_info +2024-05-22 13:30:01,081 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 13:30:01,082 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 13:30:01,130 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 13:30:01,130 INFO HandlerThread:825 [handler.py:finish():882] shutting down handler +2024-05-22 13:30:02,035 INFO WriterThread:825 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/run-pdly1svl.wandb +2024-05-22 13:30:02,096 INFO SenderThread:825 [sender.py:finish():1545] shutting down sender +2024-05-22 13:30:02,096 INFO SenderThread:825 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 13:30:02,096 INFO SenderThread:825 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..e3edb44abca59fd9e7e0403d8306d5feb6c7acba --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/logs/debug.log @@ -0,0 +1,36 @@ +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_setup.py:_flush():76] Configure stats pid to 669 +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 13:28:10,616 WARNING MainThread:669 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/logs/debug.log +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/logs/debug-internal.log +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_init.py:init():560] calling init triggers +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_init.py:init():610] starting backend +2024-05-22 13:28:10,616 INFO MainThread:669 [wandb_init.py:init():614] setting up manager +2024-05-22 13:28:10,621 INFO MainThread:669 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 13:28:10,622 INFO MainThread:669 [wandb_init.py:init():622] backend started and connected +2024-05-22 13:28:10,626 INFO MainThread:669 [wandb_init.py:init():711] updated telemetry +2024-05-22 13:28:10,634 INFO MainThread:669 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 13:28:10,926 INFO MainThread:669 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 13:28:11,041 INFO MainThread:669 [wandb_run.py:_on_init():2405] got version response +2024-05-22 13:28:11,041 INFO MainThread:669 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 13:28:11,353 INFO MainThread:669 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 13:28:11,353 INFO MainThread:669 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 13:28:11,353 INFO MainThread:669 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 13:28:11,353 INFO MainThread:669 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 13:28:11,355 INFO MainThread:669 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 13:29:54,403 INFO MainThread:669 [wandb_run.py:_config_callback():1376] config_cb None None {'task_configs': {'indiccopa-hi': {'task': 'indiccopa-hi', 'group': 'ai4bharat/IndicCOPA', 'dataset_path': 'ai4bharat/IndicCOPA', 'dataset_name': 'translation-hi', 'test_split': 'test', 'doc_to_text': "functools.partial(, connector={'cause': 'कारण', 'effect': 'परिणाम'})", 'doc_to_target': 'label', 'doc_to_choice': 'def doc_to_choice(doc):\n return [convert_choice(doc["choice1"]), convert_choice(doc["choice2"])]\n', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}}, 'cli_configs': {'model': 'hf', 'model_args': 'pretrained=/mnt/weka/peacock/idc/cronscript/ckpts//hf_ckpt//global_step100', 'batch_size': 'auto', 'batch_sizes': [64], 'device': None, 'use_cache': None, 'limit': None, 'bootstrap_iters': 100000, 'gen_kwargs': None}} +2024-05-22 13:29:55,061 INFO MainThread:669 [wandb_run.py:_finish():2103] finishing run smlgenai/bharatgpt/pdly1svl +2024-05-22 13:29:55,061 INFO MainThread:669 [wandb_run.py:_atexit_cleanup():2343] got exitcode: 0 +2024-05-22 13:29:55,061 INFO MainThread:669 [wandb_run.py:_restore():2326] restore +2024-05-22 13:29:55,061 INFO MainThread:669 [wandb_run.py:_restore():2332] restore done +2024-05-22 13:30:02,131 INFO MainThread:669 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 13:30:02,131 INFO MainThread:669 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 13:30:02,138 INFO MainThread:669 [wandb_run.py:_footer_sync_info():3953] logging synced files diff --git a/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/run-pdly1svl.wandb b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/run-pdly1svl.wandb new file mode 100644 index 0000000000000000000000000000000000000000..9677f73705e3c8458b9208957cf88cc9eda01db3 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_132810-pdly1svl/run-pdly1svl.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..28b1dbcbc34c99179201810f45483decb2e0e333 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/config.yaml @@ -0,0 +1,32 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716396406 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/output.log b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..8b137891791fe96927ad78e64b0aad7bded08bdc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/output.log @@ -0,0 +1 @@ + diff --git a/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..975ab62688eca00065a98860c51ea24f26f82f0a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T16:46:47.477861", + "startedAt": "2024-05-22T16:46:46.941325", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2326.54283125, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3380.898, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3380.921, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3372.891, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64451599121094 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..39c9344e08e1ceebd2d1da1b7477edaed6621051 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/logs/debug-internal.log @@ -0,0 +1,45 @@ +2024-05-22 16:46:46,966 INFO StreamThr :2010 [internal.py:wandb_internal():85] W&B internal server running at pid: 2010, started at: 2024-05-22 16:46:46.962262 +2024-05-22 16:46:46,969 DEBUG HandlerThread:2010 [handler.py:handle_request():158] handle_request: status +2024-05-22 16:46:46,970 INFO WriterThread:2010 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/run-s1bg4ry8.wandb +2024-05-22 16:46:46,972 DEBUG SenderThread:2010 [sender.py:send():378] send: header +2024-05-22 16:46:46,975 DEBUG SenderThread:2010 [sender.py:send():378] send: run +2024-05-22 16:46:47,249 INFO SenderThread:2010 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files +2024-05-22 16:46:47,249 INFO SenderThread:2010 [sender.py:_start_run_threads():1123] run started: s1bg4ry8 with start time 1716396406.962099 +2024-05-22 16:46:47,258 DEBUG HandlerThread:2010 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 16:46:47,258 DEBUG SenderThread:2010 [sender.py:send_request():405] send_request: check_version +2024-05-22 16:46:47,381 DEBUG HandlerThread:2010 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 16:46:47,383 DEBUG HandlerThread:2010 [system_info.py:__init__():26] System info init +2024-05-22 16:46:47,383 DEBUG HandlerThread:2010 [system_info.py:__init__():41] System info init done +2024-05-22 16:46:47,383 INFO HandlerThread:2010 [system_monitor.py:start():194] Starting system monitor +2024-05-22 16:46:47,383 INFO SystemMonitor:2010 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 16:46:47,383 INFO HandlerThread:2010 [system_monitor.py:probe():214] Collecting system info +2024-05-22 16:46:47,390 INFO SystemMonitor:2010 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 16:46:47,391 INFO SystemMonitor:2010 [interfaces.py:start():188] Started disk monitoring +2024-05-22 16:46:47,394 INFO SystemMonitor:2010 [interfaces.py:start():188] Started memory monitoring +2024-05-22 16:46:47,394 INFO SystemMonitor:2010 [interfaces.py:start():188] Started network monitoring +2024-05-22 16:46:47,477 DEBUG HandlerThread:2010 [system_info.py:probe():150] Probing system +2024-05-22 16:46:47,481 DEBUG HandlerThread:2010 [system_info.py:_probe_git():135] Probing git +2024-05-22 16:46:47,491 ERROR HandlerThread:2010 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 16:46:47,491 DEBUG HandlerThread:2010 [system_info.py:_probe_git():143] Probing git done +2024-05-22 16:46:47,491 DEBUG HandlerThread:2010 [system_info.py:probe():198] Probing system done +2024-05-22 16:46:47,491 DEBUG HandlerThread:2010 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T16:46:47.477861', 'startedAt': '2024-05-22T16:46:46.941325', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2326.54283125, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3380.898, 'min': 800.0, 'max': 3400.0}, {'current': 3380.921, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3372.891, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64451599121094}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 16:46:47,492 INFO HandlerThread:2010 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 16:46:47,492 INFO HandlerThread:2010 [system_monitor.py:probe():227] Publishing system info +2024-05-22 16:46:47,495 INFO HandlerThread:2010 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 16:46:47,500 DEBUG SenderThread:2010 [sender.py:send():378] send: files +2024-05-22 16:46:47,500 INFO SenderThread:2010 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 16:46:47,684 DEBUG HandlerThread:2010 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 16:46:47,685 DEBUG SenderThread:2010 [sender.py:send_request():405] send_request: python_packages +2024-05-22 16:46:47,686 DEBUG HandlerThread:2010 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 16:46:47,687 DEBUG SenderThread:2010 [sender.py:send_request():405] send_request: stop_status +2024-05-22 16:46:47,842 DEBUG SenderThread:2010 [sender.py:send():378] send: telemetry +2024-05-22 16:46:48,093 INFO wandb-upload_0:2010 [upload_job.py:push():130] Uploaded file /tmp/tmp8_7hxkmrwandb/g7u5619u-wandb-metadata.json +2024-05-22 16:46:48,252 INFO Thread-12 :2010 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/wandb-metadata.json +2024-05-22 16:46:48,252 INFO Thread-12 :2010 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/requirements.txt +2024-05-22 16:46:48,252 INFO Thread-12 :2010 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/output.log +2024-05-22 16:46:50,252 INFO Thread-12 :2010 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/files/output.log diff --git a/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..6605f6754ad7b248fd4c5214aeb2b8baab432da7 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/logs/debug.log @@ -0,0 +1,28 @@ +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_setup.py:_flush():76] Configure stats pid to 1855 +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 16:46:46,955 WARNING MainThread:1855 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/logs/debug.log +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/logs/debug-internal.log +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_init.py:init():560] calling init triggers +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 16:46:46,955 INFO MainThread:1855 [wandb_init.py:init():610] starting backend +2024-05-22 16:46:46,956 INFO MainThread:1855 [wandb_init.py:init():614] setting up manager +2024-05-22 16:46:46,960 INFO MainThread:1855 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 16:46:46,961 INFO MainThread:1855 [wandb_init.py:init():622] backend started and connected +2024-05-22 16:46:46,965 INFO MainThread:1855 [wandb_init.py:init():711] updated telemetry +2024-05-22 16:46:46,974 INFO MainThread:1855 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 16:46:47,258 INFO MainThread:1855 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 16:46:47,374 INFO MainThread:1855 [wandb_run.py:_on_init():2405] got version response +2024-05-22 16:46:47,374 INFO MainThread:1855 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 16:46:47,685 INFO MainThread:1855 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 16:46:47,685 INFO MainThread:1855 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 16:46:47,686 INFO MainThread:1855 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 16:46:47,686 INFO MainThread:1855 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 16:46:47,688 INFO MainThread:1855 [wandb_init.py:init():838] run started, returning control to user process diff --git a/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/run-s1bg4ry8.wandb b/lm-evaluation-harness/wandb/run-20240522_164646-s1bg4ry8/run-s1bg4ry8.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e604e579bdab9e0d3d3b7a852fab5c805be75050 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716403922 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..de348aaf46ddd51322c7eb33bb316cc722dd95cf --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log @@ -0,0 +1,34 @@ + +2024-05-22:18:52:02,888 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:18:52:11,375 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:18:52:11,376 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:18:52:11,377 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100'} +2024-05-22:18:52:13,668 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c537e3f471991310d800783f3c2d7133ee99383e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T18:52:02.677775", + "startedAt": "2024-05-22T18:52:02.112960", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.0668, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.63947296142578 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8b7cfa8286294c2bc92ff80567d26cbc6c74c4f0 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-22 18:52:02,138 INFO StreamThr :811 [internal.py:wandb_internal():85] W&B internal server running at pid: 811, started at: 2024-05-22 18:52:02.134928 +2024-05-22 18:52:02,141 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: status +2024-05-22 18:52:02,143 INFO WriterThread:811 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/run-6rpi28yk.wandb +2024-05-22 18:52:02,145 DEBUG SenderThread:811 [sender.py:send():378] send: header +2024-05-22 18:52:02,146 DEBUG SenderThread:811 [sender.py:send():378] send: run +2024-05-22 18:52:02,461 INFO SenderThread:811 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files +2024-05-22 18:52:02,461 INFO SenderThread:811 [sender.py:_start_run_threads():1123] run started: 6rpi28yk with start time 1716403922.13479 +2024-05-22 18:52:02,465 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 18:52:02,465 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: check_version +2024-05-22 18:52:02,583 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 18:52:02,585 DEBUG HandlerThread:811 [system_info.py:__init__():26] System info init +2024-05-22 18:52:02,585 DEBUG HandlerThread:811 [system_info.py:__init__():41] System info init done +2024-05-22 18:52:02,585 INFO HandlerThread:811 [system_monitor.py:start():194] Starting system monitor +2024-05-22 18:52:02,585 INFO SystemMonitor:811 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 18:52:02,585 INFO HandlerThread:811 [system_monitor.py:probe():214] Collecting system info +2024-05-22 18:52:02,592 INFO SystemMonitor:811 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 18:52:02,593 INFO SystemMonitor:811 [interfaces.py:start():188] Started disk monitoring +2024-05-22 18:52:02,594 INFO SystemMonitor:811 [interfaces.py:start():188] Started memory monitoring +2024-05-22 18:52:02,598 INFO SystemMonitor:811 [interfaces.py:start():188] Started network monitoring +2024-05-22 18:52:02,677 DEBUG HandlerThread:811 [system_info.py:probe():150] Probing system +2024-05-22 18:52:02,680 DEBUG HandlerThread:811 [system_info.py:_probe_git():135] Probing git +2024-05-22 18:52:02,690 ERROR HandlerThread:811 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 18:52:02,690 DEBUG HandlerThread:811 [system_info.py:_probe_git():143] Probing git done +2024-05-22 18:52:02,690 DEBUG HandlerThread:811 [system_info.py:probe():198] Probing system done +2024-05-22 18:52:02,690 DEBUG HandlerThread:811 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T18:52:02.677775', 'startedAt': '2024-05-22T18:52:02.112960', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.0668, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.63947296142578}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 18:52:02,691 INFO HandlerThread:811 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 18:52:02,691 INFO HandlerThread:811 [system_monitor.py:probe():227] Publishing system info +2024-05-22 18:52:02,694 INFO HandlerThread:811 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 18:52:02,699 DEBUG SenderThread:811 [sender.py:send():378] send: files +2024-05-22 18:52:02,699 INFO SenderThread:811 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 18:52:02,882 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 18:52:02,882 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: python_packages +2024-05-22 18:52:02,884 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 18:52:02,884 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: stop_status +2024-05-22 18:52:03,017 DEBUG SenderThread:811 [sender.py:send():378] send: telemetry +2024-05-22 18:52:03,289 INFO wandb-upload_0:811 [upload_job.py:push():130] Uploaded file /tmp/tmptcg2jwy8wandb/p70pxngr-wandb-metadata.json +2024-05-22 18:52:03,464 INFO Thread-12 :811 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/requirements.txt +2024-05-22 18:52:03,464 INFO Thread-12 :811 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-metadata.json +2024-05-22 18:52:03,464 INFO Thread-12 :811 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log +2024-05-22 18:52:05,464 INFO Thread-12 :811 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log +2024-05-22 18:52:08,020 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:52:13,377 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:52:13,472 INFO Thread-12 :811 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log +2024-05-22 18:52:13,676 DEBUG SenderThread:811 [sender.py:send():378] send: exit +2024-05-22 18:52:13,676 INFO SenderThread:811 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 18:52:13,676 INFO SenderThread:811 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 18:52:13,678 INFO SenderThread:811 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:52:13,678 INFO SenderThread:811 [sender.py:send_exit():593] send defer +2024-05-22 18:52:13,678 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:13,678 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 18:52:13,678 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:13,678 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 18:52:13,678 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 1 +2024-05-22 18:52:13,678 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:13,679 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 18:52:13,679 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:13,679 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 18:52:13,679 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 2 +2024-05-22 18:52:13,679 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:13,679 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 18:52:13,679 INFO HandlerThread:811 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 18:52:13,679 DEBUG SystemMonitor:811 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 18:52:13,679 DEBUG SystemMonitor:811 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 18:52:13,679 DEBUG SystemMonitor:811 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 18:52:13,682 INFO HandlerThread:811 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 18:52:13,682 INFO HandlerThread:811 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 18:52:13,682 INFO HandlerThread:811 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 18:52:13,682 INFO HandlerThread:811 [interfaces.py:finish():200] Joined network monitor +2024-05-22 18:52:13,683 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:13,683 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 18:52:13,683 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 3 +2024-05-22 18:52:13,683 DEBUG SenderThread:811 [sender.py:send():378] send: stats +2024-05-22 18:52:13,684 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:13,684 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 18:52:13,684 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:13,684 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 18:52:13,684 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 4 +2024-05-22 18:52:13,684 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:13,684 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 18:52:13,684 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:13,685 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 18:52:13,685 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 5 +2024-05-22 18:52:13,685 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:13,685 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 18:52:13,685 DEBUG SenderThread:811 [sender.py:send():378] send: summary +2024-05-22 18:52:13,686 INFO SenderThread:811 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:52:13,686 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:13,686 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 18:52:13,686 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 6 +2024-05-22 18:52:13,686 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:13,686 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 18:52:13,686 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:13,686 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 18:52:13,691 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:52:13,777 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 7 +2024-05-22 18:52:13,777 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:13,777 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 18:52:13,778 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:13,778 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 18:52:14,473 INFO Thread-12 :811 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/config.yaml +2024-05-22 18:52:14,474 INFO Thread-12 :811 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-summary.json +2024-05-22 18:52:14,676 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:52:15,042 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 8 +2024-05-22 18:52:15,042 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:52:15,042 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:15,042 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 18:52:15,043 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:15,043 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 18:52:15,043 INFO SenderThread:811 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 18:52:15,043 INFO SenderThread:811 [job_builder.py:_get_source_type():576] no source found +2024-05-22 18:52:15,043 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 9 +2024-05-22 18:52:15,043 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:15,043 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 18:52:15,044 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:15,044 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 18:52:15,044 INFO SenderThread:811 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 18:52:15,475 INFO SenderThread:811 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log +2024-05-22 18:52:15,475 INFO SenderThread:811 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files +2024-05-22 18:52:15,475 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/config.yaml config.yaml +2024-05-22 18:52:15,475 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/requirements.txt requirements.txt +2024-05-22 18:52:15,478 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log output.log +2024-05-22 18:52:15,478 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-metadata.json wandb-metadata.json +2024-05-22 18:52:15,478 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-summary.json wandb-summary.json +2024-05-22 18:52:15,478 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 10 +2024-05-22 18:52:15,478 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:15,478 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 18:52:15,479 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:15,479 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 18:52:15,479 INFO SenderThread:811 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:52:15,676 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:52:15,677 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:52:15,729 INFO wandb-upload_0:811 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/config.yaml +2024-05-22 18:52:16,109 INFO wandb-upload_3:811 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/wandb-summary.json +2024-05-22 18:52:16,119 INFO wandb-upload_2:811 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/output.log +2024-05-22 18:52:16,157 INFO wandb-upload_1:811 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/files/requirements.txt +2024-05-22 18:52:16,357 INFO Thread-11 (_thread_body):811 [sender.py:transition_state():613] send defer: 11 +2024-05-22 18:52:16,357 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:16,357 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 18:52:16,358 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:16,358 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 18:52:16,358 INFO SenderThread:811 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 18:52:16,358 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 12 +2024-05-22 18:52:16,358 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:16,358 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 18:52:16,358 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:16,358 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 18:52:16,358 INFO SenderThread:811 [file_stream.py:finish():601] file stream finish called +2024-05-22 18:52:16,422 INFO SenderThread:811 [file_stream.py:finish():605] file stream finish is done +2024-05-22 18:52:16,422 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 13 +2024-05-22 18:52:16,422 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:16,423 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 18:52:16,423 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:16,423 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 18:52:16,423 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 14 +2024-05-22 18:52:16,423 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:52:16,423 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 18:52:16,423 DEBUG SenderThread:811 [sender.py:send():378] send: final +2024-05-22 18:52:16,423 DEBUG SenderThread:811 [sender.py:send():378] send: footer +2024-05-22 18:52:16,423 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-22 18:52:16,423 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 18:52:16,424 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:52:16,424 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:52:16,424 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 18:52:16,424 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 18:52:16,424 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 18:52:16,424 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 18:52:16,425 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:52:16,425 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:52:16,425 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: server_info +2024-05-22 18:52:16,479 INFO MainThread:811 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 18:52:16,479 INFO MainThread:811 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 18:52:16,479 INFO MainThread:811 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-22 18:52:16,479 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 18:52:16,479 INFO HandlerThread:811 [handler.py:finish():882] shutting down handler +2024-05-22 18:52:17,425 INFO WriterThread:811 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/run-6rpi28yk.wandb +2024-05-22 18:52:17,479 INFO SenderThread:811 [sender.py:finish():1545] shutting down sender +2024-05-22 18:52:17,479 INFO SenderThread:811 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:52:17,479 INFO SenderThread:811 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..97aa954eba7d8e8f48637d8c2bdb2c5e2357f3a8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_setup.py:_flush():76] Configure stats pid to 655 +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 18:52:02,129 WARNING MainThread:655 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/logs/debug.log +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/logs/debug-internal.log +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_init.py:init():560] calling init triggers +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_init.py:init():610] starting backend +2024-05-22 18:52:02,129 INFO MainThread:655 [wandb_init.py:init():614] setting up manager +2024-05-22 18:52:02,133 INFO MainThread:655 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 18:52:02,134 INFO MainThread:655 [wandb_init.py:init():622] backend started and connected +2024-05-22 18:52:02,137 INFO MainThread:655 [wandb_init.py:init():711] updated telemetry +2024-05-22 18:52:02,146 INFO MainThread:655 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 18:52:02,465 INFO MainThread:655 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 18:52:02,576 INFO MainThread:655 [wandb_run.py:_on_init():2405] got version response +2024-05-22 18:52:02,576 INFO MainThread:655 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 18:52:02,883 INFO MainThread:655 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 18:52:02,883 INFO MainThread:655 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 18:52:02,883 INFO MainThread:655 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 18:52:02,883 INFO MainThread:655 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 18:52:02,885 INFO MainThread:655 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 18:52:17,481 WARNING MsgRouterThr:655 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/run-6rpi28yk.wandb b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/run-6rpi28yk.wandb new file mode 100644 index 0000000000000000000000000000000000000000..2e69b0606da5dc956c74680b44e958cdc5f57b8a Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_185202-6rpi28yk/run-6rpi28yk.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ca0b7aab65f6a636fb33f6b50a27e37867e79253 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716438199 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..65df18837893023b06388fe461506d3c02c7c223 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:04:23:20,549 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:04:23:29,093 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:04:23:29,094 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:04:23:29,094 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000'} +2024-05-23:04:23:31,427 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..88181d04cb90f3bd8f00a85cc517ce4f45bd5aed --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/requirements.txt @@ -0,0 +1,156 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +megatron-lm==1.1.5 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..ed043b9f05b1d94ee80c075076a508e63cc2158c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T04:23:20.338217", + "startedAt": "2024-05-23T04:23:19.803606", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.5000124999997, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2906.271, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.6305694580078 + } + }, + "memory": { + "total": 1007.4380111694336 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..b04b8862168f299e0b9f5400800df75a26ce7b2b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 04:23:19,824 INFO StreamThr :1064 [internal.py:wandb_internal():85] W&B internal server running at pid: 1064, started at: 2024-05-23 04:23:19.822683 +2024-05-23 04:23:19,829 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: status +2024-05-23 04:23:19,830 INFO WriterThread:1064 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/run-9d6pz5i2.wandb +2024-05-23 04:23:19,833 DEBUG SenderThread:1064 [sender.py:send():378] send: header +2024-05-23 04:23:19,835 DEBUG SenderThread:1064 [sender.py:send():378] send: run +2024-05-23 04:23:20,138 INFO SenderThread:1064 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files +2024-05-23 04:23:20,138 INFO SenderThread:1064 [sender.py:_start_run_threads():1123] run started: 9d6pz5i2 with start time 1716438199.822539 +2024-05-23 04:23:20,143 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 04:23:20,143 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: check_version +2024-05-23 04:23:20,264 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 04:23:20,266 DEBUG HandlerThread:1064 [system_info.py:__init__():26] System info init +2024-05-23 04:23:20,267 DEBUG HandlerThread:1064 [system_info.py:__init__():41] System info init done +2024-05-23 04:23:20,267 INFO HandlerThread:1064 [system_monitor.py:start():194] Starting system monitor +2024-05-23 04:23:20,267 INFO SystemMonitor:1064 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 04:23:20,267 INFO HandlerThread:1064 [system_monitor.py:probe():214] Collecting system info +2024-05-23 04:23:20,273 INFO SystemMonitor:1064 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 04:23:20,274 INFO SystemMonitor:1064 [interfaces.py:start():188] Started disk monitoring +2024-05-23 04:23:20,279 INFO SystemMonitor:1064 [interfaces.py:start():188] Started memory monitoring +2024-05-23 04:23:20,280 INFO SystemMonitor:1064 [interfaces.py:start():188] Started network monitoring +2024-05-23 04:23:20,338 DEBUG HandlerThread:1064 [system_info.py:probe():150] Probing system +2024-05-23 04:23:20,341 DEBUG HandlerThread:1064 [system_info.py:_probe_git():135] Probing git +2024-05-23 04:23:20,352 ERROR HandlerThread:1064 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 04:23:20,352 DEBUG HandlerThread:1064 [system_info.py:_probe_git():143] Probing git done +2024-05-23 04:23:20,352 DEBUG HandlerThread:1064 [system_info.py:probe():198] Probing system done +2024-05-23 04:23:20,352 DEBUG HandlerThread:1064 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T04:23:20.338217', 'startedAt': '2024-05-23T04:23:19.803606', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.5000124999997, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2906.271, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.6305694580078}}, 'memory': {'total': 1007.4380111694336}} +2024-05-23 04:23:20,352 INFO HandlerThread:1064 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 04:23:20,352 INFO HandlerThread:1064 [system_monitor.py:probe():227] Publishing system info +2024-05-23 04:23:20,355 INFO HandlerThread:1064 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 04:23:20,361 DEBUG SenderThread:1064 [sender.py:send():378] send: files +2024-05-23 04:23:20,361 INFO SenderThread:1064 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 04:23:20,541 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 04:23:20,541 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: python_packages +2024-05-23 04:23:20,543 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 04:23:20,543 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: stop_status +2024-05-23 04:23:20,680 DEBUG SenderThread:1064 [sender.py:send():378] send: telemetry +2024-05-23 04:23:21,007 INFO wandb-upload_0:1064 [upload_job.py:push():130] Uploaded file /tmp/tmp6fyl7pz3wandb/gpskbk2d-wandb-metadata.json +2024-05-23 04:23:21,141 INFO Thread-12 :1064 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/requirements.txt +2024-05-23 04:23:21,141 INFO Thread-12 :1064 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log +2024-05-23 04:23:21,142 INFO Thread-12 :1064 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-metadata.json +2024-05-23 04:23:23,141 INFO Thread-12 :1064 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log +2024-05-23 04:23:25,687 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 04:23:31,095 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 04:23:31,151 INFO Thread-12 :1064 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log +2024-05-23 04:23:31,435 DEBUG SenderThread:1064 [sender.py:send():378] send: exit +2024-05-23 04:23:31,435 INFO SenderThread:1064 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 04:23:31,435 INFO SenderThread:1064 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 04:23:31,437 INFO SenderThread:1064 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 04:23:31,437 INFO SenderThread:1064 [sender.py:send_exit():593] send defer +2024-05-23 04:23:31,437 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:31,437 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 04:23:31,437 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:31,437 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 04:23:31,437 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 1 +2024-05-23 04:23:31,437 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:31,437 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 04:23:31,438 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:31,438 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 04:23:31,438 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 2 +2024-05-23 04:23:31,438 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:31,438 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 04:23:31,438 INFO HandlerThread:1064 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 04:23:31,438 DEBUG SystemMonitor:1064 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 04:23:31,438 DEBUG SystemMonitor:1064 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 04:23:31,438 DEBUG SystemMonitor:1064 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 04:23:31,441 INFO HandlerThread:1064 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 04:23:31,441 INFO HandlerThread:1064 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 04:23:31,441 INFO HandlerThread:1064 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 04:23:31,441 INFO HandlerThread:1064 [interfaces.py:finish():200] Joined network monitor +2024-05-23 04:23:31,442 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:31,442 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 04:23:31,442 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 3 +2024-05-23 04:23:31,442 DEBUG SenderThread:1064 [sender.py:send():378] send: stats +2024-05-23 04:23:31,443 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:31,443 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 04:23:31,444 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:31,444 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 04:23:31,444 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 4 +2024-05-23 04:23:31,444 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:31,444 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 04:23:31,444 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:31,444 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 04:23:31,444 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 5 +2024-05-23 04:23:31,444 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:31,444 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 04:23:31,444 DEBUG SenderThread:1064 [sender.py:send():378] send: summary +2024-05-23 04:23:31,445 INFO SenderThread:1064 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 04:23:31,445 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:31,445 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 04:23:31,445 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 6 +2024-05-23 04:23:31,445 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:31,446 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 04:23:31,446 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:31,446 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 04:23:31,450 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 04:23:31,522 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 7 +2024-05-23 04:23:31,523 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:31,523 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 04:23:31,523 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:31,523 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 04:23:32,153 INFO Thread-12 :1064 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/config.yaml +2024-05-23 04:23:32,153 INFO Thread-12 :1064 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-summary.json +2024-05-23 04:23:32,435 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 04:23:32,708 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 8 +2024-05-23 04:23:32,708 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 04:23:32,708 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:32,709 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 04:23:32,709 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:32,709 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 04:23:32,709 INFO SenderThread:1064 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 04:23:32,709 INFO SenderThread:1064 [job_builder.py:_get_source_type():576] no source found +2024-05-23 04:23:32,710 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 9 +2024-05-23 04:23:32,710 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:32,710 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 04:23:32,710 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:32,710 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 04:23:32,710 INFO SenderThread:1064 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 04:23:33,154 INFO SenderThread:1064 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log +2024-05-23 04:23:33,154 INFO SenderThread:1064 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files +2024-05-23 04:23:33,155 INFO SenderThread:1064 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-metadata.json wandb-metadata.json +2024-05-23 04:23:33,155 INFO SenderThread:1064 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/requirements.txt requirements.txt +2024-05-23 04:23:33,155 INFO SenderThread:1064 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/config.yaml config.yaml +2024-05-23 04:23:33,157 INFO SenderThread:1064 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-summary.json wandb-summary.json +2024-05-23 04:23:33,159 INFO SenderThread:1064 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log output.log +2024-05-23 04:23:33,161 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 10 +2024-05-23 04:23:33,162 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:33,162 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 04:23:33,162 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:33,164 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 04:23:33,164 INFO SenderThread:1064 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 04:23:33,432 INFO wandb-upload_0:1064 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/requirements.txt +2024-05-23 04:23:33,435 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 04:23:33,442 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 04:23:33,747 INFO wandb-upload_1:1064 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/config.yaml +2024-05-23 04:23:33,776 INFO wandb-upload_3:1064 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/output.log +2024-05-23 04:23:33,791 INFO wandb-upload_2:1064 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/files/wandb-summary.json +2024-05-23 04:23:33,992 INFO Thread-11 (_thread_body):1064 [sender.py:transition_state():613] send defer: 11 +2024-05-23 04:23:33,992 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:33,992 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 04:23:33,992 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:33,992 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 04:23:33,992 INFO SenderThread:1064 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 04:23:33,992 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 12 +2024-05-23 04:23:33,992 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:33,992 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 04:23:33,993 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:33,993 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 04:23:33,993 INFO SenderThread:1064 [file_stream.py:finish():601] file stream finish called +2024-05-23 04:23:34,126 INFO SenderThread:1064 [file_stream.py:finish():605] file stream finish is done +2024-05-23 04:23:34,126 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 13 +2024-05-23 04:23:34,126 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:34,126 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 04:23:34,126 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:34,126 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 04:23:34,126 INFO SenderThread:1064 [sender.py:transition_state():613] send defer: 14 +2024-05-23 04:23:34,126 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:23:34,126 INFO HandlerThread:1064 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 04:23:34,127 DEBUG SenderThread:1064 [sender.py:send():378] send: final +2024-05-23 04:23:34,127 DEBUG SenderThread:1064 [sender.py:send():378] send: footer +2024-05-23 04:23:34,127 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: defer +2024-05-23 04:23:34,127 INFO SenderThread:1064 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 04:23:34,127 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 04:23:34,127 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 04:23:34,128 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 04:23:34,128 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 04:23:34,128 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 04:23:34,128 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 04:23:34,128 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 04:23:34,128 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 04:23:34,128 DEBUG SenderThread:1064 [sender.py:send_request():405] send_request: server_info +2024-05-23 04:23:34,181 INFO MainThread:1064 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 04:23:34,181 INFO MainThread:1064 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 04:23:34,181 INFO MainThread:1064 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 04:23:34,181 DEBUG HandlerThread:1064 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 04:23:34,181 INFO HandlerThread:1064 [handler.py:finish():882] shutting down handler +2024-05-23 04:23:35,128 INFO WriterThread:1064 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/run-9d6pz5i2.wandb +2024-05-23 04:23:35,181 INFO SenderThread:1064 [sender.py:finish():1545] shutting down sender +2024-05-23 04:23:35,181 INFO SenderThread:1064 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 04:23:35,181 INFO SenderThread:1064 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..ecb904f1aebbe83c3a8d75b10e18a1b2eeb10b91 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 04:23:19,817 INFO MainThread:909 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 04:23:19,817 INFO MainThread:909 [wandb_setup.py:_flush():76] Configure stats pid to 909 +2024-05-23 04:23:19,817 INFO MainThread:909 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 04:23:19,817 INFO MainThread:909 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 04:23:19,817 INFO MainThread:909 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 04:23:19,817 INFO MainThread:909 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 04:23:19,817 WARNING MainThread:909 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 04:23:19,817 INFO MainThread:909 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 04:23:19,817 INFO MainThread:909 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 04:23:19,818 INFO MainThread:909 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/logs/debug.log +2024-05-23 04:23:19,818 INFO MainThread:909 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/logs/debug-internal.log +2024-05-23 04:23:19,818 INFO MainThread:909 [wandb_init.py:init():560] calling init triggers +2024-05-23 04:23:19,818 INFO MainThread:909 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 04:23:19,818 INFO MainThread:909 [wandb_init.py:init():610] starting backend +2024-05-23 04:23:19,818 INFO MainThread:909 [wandb_init.py:init():614] setting up manager +2024-05-23 04:23:19,821 INFO MainThread:909 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 04:23:19,822 INFO MainThread:909 [wandb_init.py:init():622] backend started and connected +2024-05-23 04:23:19,826 INFO MainThread:909 [wandb_init.py:init():711] updated telemetry +2024-05-23 04:23:19,835 INFO MainThread:909 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 04:23:20,143 INFO MainThread:909 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 04:23:20,258 INFO MainThread:909 [wandb_run.py:_on_init():2405] got version response +2024-05-23 04:23:20,258 INFO MainThread:909 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 04:23:20,542 INFO MainThread:909 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 04:23:20,542 INFO MainThread:909 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 04:23:20,542 INFO MainThread:909 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 04:23:20,542 INFO MainThread:909 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 04:23:20,546 INFO MainThread:909 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 04:23:35,182 WARNING MsgRouterThr:909 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/run-9d6pz5i2.wandb b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/run-9d6pz5i2.wandb new file mode 100644 index 0000000000000000000000000000000000000000..cdde1260026c0e3e098e4249d5d0e099cd036bfa Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_042319-9d6pz5i2/run-9d6pz5i2.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..f675c3016b5332c1acf28f436e0b60adeead9c12 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.0 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..3c26e673f0a1491e966b77b5908d47160b070941 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 13:08:12,722 INFO StreamThr :2871 [internal.py:wandb_internal():85] W&B internal server running at pid: 2871, started at: 2024-05-23 13:08:12.720746 +2024-05-23 13:08:12,727 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: status +2024-05-23 13:08:12,727 INFO WriterThread:2871 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/run-gglzto8s.wandb +2024-05-23 13:08:12,729 DEBUG SenderThread:2871 [sender.py:send():378] send: header +2024-05-23 13:08:12,732 DEBUG SenderThread:2871 [sender.py:send():378] send: run +2024-05-23 13:08:13,052 INFO SenderThread:2871 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files +2024-05-23 13:08:13,053 INFO SenderThread:2871 [sender.py:_start_run_threads():1123] run started: gglzto8s with start time 1716469692.720604 +2024-05-23 13:08:13,053 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 13:08:13,054 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: check_version +2024-05-23 13:08:13,174 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 13:08:13,176 DEBUG HandlerThread:2871 [system_info.py:__init__():26] System info init +2024-05-23 13:08:13,176 DEBUG HandlerThread:2871 [system_info.py:__init__():41] System info init done +2024-05-23 13:08:13,176 INFO HandlerThread:2871 [system_monitor.py:start():194] Starting system monitor +2024-05-23 13:08:13,176 INFO SystemMonitor:2871 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 13:08:13,176 INFO HandlerThread:2871 [system_monitor.py:probe():214] Collecting system info +2024-05-23 13:08:13,183 INFO SystemMonitor:2871 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 13:08:13,183 INFO SystemMonitor:2871 [interfaces.py:start():188] Started disk monitoring +2024-05-23 13:08:13,184 INFO SystemMonitor:2871 [interfaces.py:start():188] Started memory monitoring +2024-05-23 13:08:13,184 INFO SystemMonitor:2871 [interfaces.py:start():188] Started network monitoring +2024-05-23 13:08:13,268 DEBUG HandlerThread:2871 [system_info.py:probe():150] Probing system +2024-05-23 13:08:13,272 DEBUG HandlerThread:2871 [system_info.py:_probe_git():135] Probing git +2024-05-23 13:08:13,282 ERROR HandlerThread:2871 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 13:08:13,282 DEBUG HandlerThread:2871 [system_info.py:_probe_git():143] Probing git done +2024-05-23 13:08:13,282 DEBUG HandlerThread:2871 [system_info.py:probe():198] Probing system done +2024-05-23 13:08:13,282 DEBUG HandlerThread:2871 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T13:08:13.268953', 'startedAt': '2024-05-23T13:08:12.700786', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.5000124999997, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.62897872924805}}, 'memory': {'total': 1007.4379539489746}} +2024-05-23 13:08:13,282 INFO HandlerThread:2871 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 13:08:13,282 INFO HandlerThread:2871 [system_monitor.py:probe():227] Publishing system info +2024-05-23 13:08:13,285 INFO HandlerThread:2871 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 13:08:13,290 DEBUG SenderThread:2871 [sender.py:send():378] send: files +2024-05-23 13:08:13,290 INFO SenderThread:2871 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 13:08:13,469 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 13:08:13,469 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: python_packages +2024-05-23 13:08:13,471 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 13:08:13,471 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: stop_status +2024-05-23 13:08:13,590 DEBUG SenderThread:2871 [sender.py:send():378] send: telemetry +2024-05-23 13:08:13,849 INFO wandb-upload_0:2871 [upload_job.py:push():130] Uploaded file /tmp/tmpy_8is2jlwandb/ismmeup1-wandb-metadata.json +2024-05-23 13:08:14,054 INFO Thread-12 :2871 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/requirements.txt +2024-05-23 13:08:14,054 INFO Thread-12 :2871 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/wandb-metadata.json +2024-05-23 13:08:14,054 INFO Thread-12 :2871 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/output.log +2024-05-23 13:08:16,053 INFO Thread-12 :2871 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/output.log +2024-05-23 13:08:18,592 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:08:23,944 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:08:24,062 INFO Thread-12 :2871 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/output.log +2024-05-23 13:08:24,242 DEBUG SenderThread:2871 [sender.py:send():378] send: exit +2024-05-23 13:08:24,243 INFO SenderThread:2871 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 13:08:24,243 INFO SenderThread:2871 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 13:08:24,244 INFO SenderThread:2871 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:08:24,244 INFO SenderThread:2871 [sender.py:send_exit():593] send defer +2024-05-23 13:08:24,244 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:24,244 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 13:08:24,245 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:24,245 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 13:08:24,245 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 1 +2024-05-23 13:08:24,245 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:24,245 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 13:08:24,245 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:24,245 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 13:08:24,245 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 2 +2024-05-23 13:08:24,245 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:24,245 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 13:08:24,245 INFO HandlerThread:2871 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 13:08:24,245 DEBUG SystemMonitor:2871 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 13:08:24,245 DEBUG SystemMonitor:2871 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 13:08:24,246 DEBUG SystemMonitor:2871 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 13:08:24,246 INFO HandlerThread:2871 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 13:08:24,249 INFO HandlerThread:2871 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 13:08:24,249 INFO HandlerThread:2871 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 13:08:24,249 INFO HandlerThread:2871 [interfaces.py:finish():200] Joined network monitor +2024-05-23 13:08:24,249 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:24,249 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 13:08:24,249 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 3 +2024-05-23 13:08:24,250 DEBUG SenderThread:2871 [sender.py:send():378] send: stats +2024-05-23 13:08:24,251 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:24,251 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 13:08:24,251 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:24,251 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 13:08:24,251 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 4 +2024-05-23 13:08:24,251 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:24,251 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 13:08:24,251 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:24,251 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 13:08:24,251 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 5 +2024-05-23 13:08:24,251 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:24,251 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 13:08:24,251 DEBUG SenderThread:2871 [sender.py:send():378] send: summary +2024-05-23 13:08:24,252 INFO SenderThread:2871 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 13:08:24,252 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:24,252 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 13:08:24,252 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 6 +2024-05-23 13:08:24,252 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:24,253 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 13:08:24,253 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:24,253 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 13:08:24,257 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 13:08:24,345 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 7 +2024-05-23 13:08:24,345 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:24,345 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 13:08:24,345 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:24,345 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 13:08:25,063 INFO Thread-12 :2871 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/config.yaml +2024-05-23 13:08:25,063 INFO Thread-12 :2871 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/wandb-summary.json +2024-05-23 13:08:25,243 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:08:25,612 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 8 +2024-05-23 13:08:25,612 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:08:25,612 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:25,612 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 13:08:25,612 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:25,612 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 13:08:25,612 INFO SenderThread:2871 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 13:08:25,613 INFO SenderThread:2871 [job_builder.py:_get_source_type():576] no source found +2024-05-23 13:08:25,613 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 9 +2024-05-23 13:08:25,613 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:25,613 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 13:08:25,613 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:25,613 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 13:08:25,613 INFO SenderThread:2871 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 13:08:26,065 INFO SenderThread:2871 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/output.log +2024-05-23 13:08:26,065 INFO SenderThread:2871 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files +2024-05-23 13:08:26,065 INFO SenderThread:2871 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/requirements.txt requirements.txt +2024-05-23 13:08:26,065 INFO SenderThread:2871 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/output.log output.log +2024-05-23 13:08:26,068 INFO SenderThread:2871 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/config.yaml config.yaml +2024-05-23 13:08:26,068 INFO SenderThread:2871 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/wandb-metadata.json wandb-metadata.json +2024-05-23 13:08:26,068 INFO SenderThread:2871 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/wandb-summary.json wandb-summary.json +2024-05-23 13:08:26,068 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 10 +2024-05-23 13:08:26,068 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:26,068 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 13:08:26,069 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:26,069 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 13:08:26,069 INFO SenderThread:2871 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:08:26,243 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:08:26,243 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:08:26,433 INFO wandb-upload_0:2871 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/requirements.txt +2024-05-23 13:08:26,682 INFO wandb-upload_2:2871 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/config.yaml +2024-05-23 13:08:26,685 INFO wandb-upload_3:2871 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/wandb-summary.json +2024-05-23 13:08:26,695 INFO wandb-upload_1:2871 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/files/output.log +2024-05-23 13:08:26,895 INFO Thread-11 (_thread_body):2871 [sender.py:transition_state():613] send defer: 11 +2024-05-23 13:08:26,895 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:26,895 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 13:08:26,895 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:26,896 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 13:08:26,896 INFO SenderThread:2871 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 13:08:26,896 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 12 +2024-05-23 13:08:26,896 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:26,896 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 13:08:26,896 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:26,896 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 13:08:26,896 INFO SenderThread:2871 [file_stream.py:finish():601] file stream finish called +2024-05-23 13:08:26,974 INFO SenderThread:2871 [file_stream.py:finish():605] file stream finish is done +2024-05-23 13:08:26,974 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 13 +2024-05-23 13:08:26,974 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:26,974 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 13:08:26,974 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:26,974 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 13:08:26,974 INFO SenderThread:2871 [sender.py:transition_state():613] send defer: 14 +2024-05-23 13:08:26,974 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: defer +2024-05-23 13:08:26,974 INFO HandlerThread:2871 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 13:08:26,974 DEBUG SenderThread:2871 [sender.py:send():378] send: final +2024-05-23 13:08:26,975 DEBUG SenderThread:2871 [sender.py:send():378] send: footer +2024-05-23 13:08:26,975 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: defer +2024-05-23 13:08:26,975 INFO SenderThread:2871 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 13:08:26,975 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:08:26,975 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:08:26,976 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 13:08:26,976 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 13:08:26,976 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 13:08:26,976 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 13:08:26,976 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 13:08:26,976 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 13:08:26,976 DEBUG SenderThread:2871 [sender.py:send_request():405] send_request: server_info +2024-05-23 13:08:27,031 INFO MainThread:2871 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 13:08:27,031 INFO MainThread:2871 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 13:08:27,031 INFO MainThread:2871 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 13:08:27,031 DEBUG HandlerThread:2871 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 13:08:27,031 INFO HandlerThread:2871 [handler.py:finish():882] shutting down handler +2024-05-23 13:08:27,976 INFO WriterThread:2871 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/run-gglzto8s.wandb +2024-05-23 13:08:28,031 INFO SenderThread:2871 [sender.py:finish():1545] shutting down sender +2024-05-23 13:08:28,031 INFO SenderThread:2871 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 13:08:28,031 INFO SenderThread:2871 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..629d27329b897d7bcc0e8ab58b960ea14d046dcc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 13:08:12,715 INFO MainThread:2716 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 13:08:12,715 INFO MainThread:2716 [wandb_setup.py:_flush():76] Configure stats pid to 2716 +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 13:08:12,716 WARNING MainThread:2716 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/logs/debug.log +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/logs/debug-internal.log +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_init.py:init():560] calling init triggers +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_init.py:init():610] starting backend +2024-05-23 13:08:12,716 INFO MainThread:2716 [wandb_init.py:init():614] setting up manager +2024-05-23 13:08:12,719 INFO MainThread:2716 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 13:08:12,720 INFO MainThread:2716 [wandb_init.py:init():622] backend started and connected +2024-05-23 13:08:12,723 INFO MainThread:2716 [wandb_init.py:init():711] updated telemetry +2024-05-23 13:08:12,731 INFO MainThread:2716 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 13:08:13,053 INFO MainThread:2716 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 13:08:13,168 INFO MainThread:2716 [wandb_run.py:_on_init():2405] got version response +2024-05-23 13:08:13,168 INFO MainThread:2716 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 13:08:13,470 INFO MainThread:2716 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 13:08:13,470 INFO MainThread:2716 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 13:08:13,470 INFO MainThread:2716 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 13:08:13,470 INFO MainThread:2716 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 13:08:13,473 INFO MainThread:2716 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 13:08:28,032 WARNING MsgRouterThr:2716 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/run-gglzto8s.wandb b/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/run-gglzto8s.wandb new file mode 100644 index 0000000000000000000000000000000000000000..35632cbdff4956a56c082a34b159ca6df3dad906 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_130812-gglzto8s/run-gglzto8s.wandb differ diff --git a/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/INSTALLER b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/INSTALLER new file mode 100644 index 0000000000000000000000000000000000000000..a1b589e38a32041e49332e5e81c2d363dc418d68 --- /dev/null +++ b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/INSTALLER @@ -0,0 +1 @@ +pip diff --git a/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/LICENSE b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/LICENSE new file mode 100644 index 0000000000000000000000000000000000000000..2f1b8e15e5627d92f0521605c9870bc8e5505cb4 --- /dev/null +++ b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/LICENSE @@ -0,0 +1,20 @@ +Copyright (c) 2017-2021 Ingy döt Net +Copyright (c) 2006-2016 Kirill Simonov + +Permission is hereby granted, free of charge, to any person obtaining a copy of +this software and associated documentation files (the "Software"), to deal in +the Software without restriction, including without limitation the rights to +use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies +of the Software, and to permit persons to whom the Software is furnished to do +so, subject to the following conditions: + +The above copyright notice and this permission notice shall be included in all +copies or substantial portions of the Software. + +THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR +IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, +FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE +AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER +LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE +SOFTWARE. diff --git a/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/METADATA b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/METADATA new file mode 100644 index 0000000000000000000000000000000000000000..c8905983e369893f68879f4cdfb7290d54d5f822 --- /dev/null +++ b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/METADATA @@ -0,0 +1,46 @@ +Metadata-Version: 2.1 +Name: PyYAML +Version: 6.0.1 +Summary: YAML parser and emitter for Python +Home-page: https://pyyaml.org/ +Download-URL: https://pypi.org/project/PyYAML/ +Author: Kirill Simonov +Author-email: xi@resolvent.net +License: MIT +Project-URL: Bug Tracker, https://github.com/yaml/pyyaml/issues +Project-URL: CI, https://github.com/yaml/pyyaml/actions +Project-URL: Documentation, https://pyyaml.org/wiki/PyYAMLDocumentation +Project-URL: Mailing lists, http://lists.sourceforge.net/lists/listinfo/yaml-core +Project-URL: Source Code, https://github.com/yaml/pyyaml +Platform: Any +Classifier: Development Status :: 5 - Production/Stable +Classifier: Intended Audience :: Developers +Classifier: License :: OSI Approved :: MIT License +Classifier: Operating System :: OS Independent +Classifier: Programming Language :: Cython +Classifier: Programming Language :: Python +Classifier: Programming Language :: Python :: 3 +Classifier: Programming Language :: Python :: 3.6 +Classifier: Programming Language :: Python :: 3.7 +Classifier: Programming Language :: Python :: 3.8 +Classifier: Programming Language :: Python :: 3.9 +Classifier: Programming Language :: Python :: 3.10 +Classifier: Programming Language :: Python :: 3.11 +Classifier: Programming Language :: Python :: Implementation :: CPython +Classifier: Programming Language :: Python :: Implementation :: PyPy +Classifier: Topic :: Software Development :: Libraries :: Python Modules +Classifier: Topic :: Text Processing :: Markup +Requires-Python: >=3.6 +License-File: LICENSE + +YAML is a data serialization format designed for human readability +and interaction with scripting languages. PyYAML is a YAML parser +and emitter for Python. + +PyYAML features a complete YAML 1.1 parser, Unicode support, pickle +support, capable extension API, and sensible error messages. PyYAML +supports standard YAML tags and provides Python-specific tags that +allow to represent an arbitrary Python object. + +PyYAML is applicable for a broad range of tasks from complex +configuration files to object serialization and persistence. diff --git a/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/RECORD b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/RECORD new file mode 100644 index 0000000000000000000000000000000000000000..90aeed85f892d7de7ae8ffad5477859e10a5cde1 --- /dev/null +++ b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/RECORD @@ -0,0 +1,43 @@ +PyYAML-6.0.1.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4 +PyYAML-6.0.1.dist-info/LICENSE,sha256=jTko-dxEkP1jVwfLiOsmvXZBAqcoKVQwfT5RZ6V36KQ,1101 +PyYAML-6.0.1.dist-info/METADATA,sha256=UNNF8-SzzwOKXVo-kV5lXUGH2_wDWMBmGxqISpp5HQk,2058 +PyYAML-6.0.1.dist-info/RECORD,, +PyYAML-6.0.1.dist-info/WHEEL,sha256=iZaXX0Td62Nww8bojl0E84uJHjT41csHPKZmbUBbJPs,152 +PyYAML-6.0.1.dist-info/top_level.txt,sha256=rpj0IVMTisAjh_1vG3Ccf9v5jpCQwAz6cD1IVU5ZdhQ,11 +_yaml/__init__.py,sha256=04Ae_5osxahpJHa3XBZUAf4wi6XX32gR8D6X6p64GEA,1402 +_yaml/__pycache__/__init__.cpython-310.pyc,, +yaml/__init__.py,sha256=bhl05qSeO-1ZxlSRjGrvl2m9nrXb1n9-GQatTN0Mrqc,12311 +yaml/__pycache__/__init__.cpython-310.pyc,, +yaml/__pycache__/composer.cpython-310.pyc,, +yaml/__pycache__/constructor.cpython-310.pyc,, +yaml/__pycache__/cyaml.cpython-310.pyc,, +yaml/__pycache__/dumper.cpython-310.pyc,, +yaml/__pycache__/emitter.cpython-310.pyc,, +yaml/__pycache__/error.cpython-310.pyc,, +yaml/__pycache__/events.cpython-310.pyc,, +yaml/__pycache__/loader.cpython-310.pyc,, +yaml/__pycache__/nodes.cpython-310.pyc,, +yaml/__pycache__/parser.cpython-310.pyc,, +yaml/__pycache__/reader.cpython-310.pyc,, +yaml/__pycache__/representer.cpython-310.pyc,, +yaml/__pycache__/resolver.cpython-310.pyc,, +yaml/__pycache__/scanner.cpython-310.pyc,, +yaml/__pycache__/serializer.cpython-310.pyc,, +yaml/__pycache__/tokens.cpython-310.pyc,, +yaml/_yaml.cpython-310-x86_64-linux-gnu.so,sha256=_9iVrASatQgQSFXlKeCe2uK2TyKwk9nd61Cs_-fqAHM,2226000 +yaml/composer.py,sha256=_Ko30Wr6eDWUeUpauUGT3Lcg9QPBnOPVlTnIMRGJ9FM,4883 +yaml/constructor.py,sha256=kNgkfaeLUkwQYY_Q6Ff1Tz2XVw_pG1xVE9Ak7z-viLA,28639 +yaml/cyaml.py,sha256=6ZrAG9fAYvdVe2FK_w0hmXoG7ZYsoYUwapG8CiC72H0,3851 +yaml/dumper.py,sha256=PLctZlYwZLp7XmeUdwRuv4nYOZ2UBnDIUy8-lKfLF-o,2837 +yaml/emitter.py,sha256=jghtaU7eFwg31bG0B7RZea_29Adi9CKmXq_QjgQpCkQ,43006 +yaml/error.py,sha256=Ah9z-toHJUbE9j-M8YpxgSRM5CgLCcwVzJgLLRF2Fxo,2533 +yaml/events.py,sha256=50_TksgQiE4up-lKo_V-nBy-tAIxkIPQxY5qDhKCeHw,2445 +yaml/loader.py,sha256=UVa-zIqmkFSCIYq_PgSGm4NSJttHY2Rf_zQ4_b1fHN0,2061 +yaml/nodes.py,sha256=gPKNj8pKCdh2d4gr3gIYINnPOaOxGhJAUiYhGRnPE84,1440 +yaml/parser.py,sha256=ilWp5vvgoHFGzvOZDItFoGjD6D42nhlZrZyjAwa0oJo,25495 +yaml/reader.py,sha256=0dmzirOiDG4Xo41RnuQS7K9rkY3xjHiVasfDMNTqCNw,6794 +yaml/representer.py,sha256=IuWP-cAW9sHKEnS0gCqSa894k1Bg4cgTxaDwIcbRQ-Y,14190 +yaml/resolver.py,sha256=9L-VYfm4mWHxUD1Vg4X7rjDRK_7VZd6b92wzq7Y2IKY,9004 +yaml/scanner.py,sha256=YEM3iLZSaQwXcQRg2l2R4MdT0zGP2F9eHkKGKnHyWQY,51279 +yaml/serializer.py,sha256=ChuFgmhU01hj4xgI8GaKv6vfM2Bujwa9i7d2FAHj7cA,4165 +yaml/tokens.py,sha256=lTQIzSVw8Mg9wv459-TjiOQe6wVziqaRlqX2_89rp54,2573 diff --git a/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/WHEEL b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/WHEEL new file mode 100644 index 0000000000000000000000000000000000000000..2d1b4b87130cdd6e713250ed45eec87c87d0405f --- /dev/null +++ b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/WHEEL @@ -0,0 +1,6 @@ +Wheel-Version: 1.0 +Generator: bdist_wheel (0.40.0) +Root-Is-Purelib: false +Tag: cp310-cp310-manylinux_2_17_x86_64 +Tag: cp310-cp310-manylinux2014_x86_64 + diff --git a/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/top_level.txt b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/top_level.txt new file mode 100644 index 0000000000000000000000000000000000000000..e6475e911f628412049bc4090d86f23ac403adde --- /dev/null +++ b/venv/lib/python3.10/site-packages/PyYAML-6.0.1.dist-info/top_level.txt @@ -0,0 +1,2 @@ +_yaml +yaml diff --git a/venv/lib/python3.10/site-packages/absl/__init__.py b/venv/lib/python3.10/site-packages/absl/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..a3bd1cd51810385ca0e5e9fed3fb9a804febf27e --- /dev/null +++ b/venv/lib/python3.10/site-packages/absl/__init__.py @@ -0,0 +1,13 @@ +# Copyright 2017 The Abseil Authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. diff --git a/venv/lib/python3.10/site-packages/absl/app.py b/venv/lib/python3.10/site-packages/absl/app.py new file mode 100644 index 0000000000000000000000000000000000000000..d12397b31a9093dd0f9a8622c72a16d31fcb4fa9 --- /dev/null +++ b/venv/lib/python3.10/site-packages/absl/app.py @@ -0,0 +1,480 @@ +# Copyright 2017 The Abseil Authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +"""Generic entry point for Abseil Python applications. + +To use this module, define a ``main`` function with a single ``argv`` argument +and call ``app.run(main)``. For example:: + + def main(argv): + if len(argv) > 1: + raise app.UsageError('Too many command-line arguments.') + + if __name__ == '__main__': + app.run(main) +""" + +import collections +import errno +import os +import pdb +import sys +import textwrap +import traceback + +from absl import command_name +from absl import flags +from absl import logging + +try: + import faulthandler +except ImportError: + faulthandler = None + +FLAGS = flags.FLAGS + +flags.DEFINE_boolean('run_with_pdb', False, 'Set to true for PDB debug mode') +flags.DEFINE_boolean('pdb_post_mortem', False, + 'Set to true to handle uncaught exceptions with PDB ' + 'post mortem.') +flags.DEFINE_alias('pdb', 'pdb_post_mortem') +flags.DEFINE_boolean('run_with_profiling', False, + 'Set to true for profiling the script. ' + 'Execution will be slower, and the output format might ' + 'change over time.') +flags.DEFINE_string('profile_file', None, + 'Dump profile information to a file (for python -m ' + 'pstats). Implies --run_with_profiling.') +flags.DEFINE_boolean('use_cprofile_for_profiling', True, + 'Use cProfile instead of the profile module for ' + 'profiling. This has no effect unless ' + '--run_with_profiling is set.') +flags.DEFINE_boolean('only_check_args', False, + 'Set to true to validate args and exit.', + allow_hide_cpp=True) + + +# If main() exits via an abnormal exception, call into these +# handlers before exiting. +EXCEPTION_HANDLERS = [] + + +class Error(Exception): + pass + + +class UsageError(Error): + """Exception raised when the arguments supplied by the user are invalid. + + Raise this when the arguments supplied are invalid from the point of + view of the application. For example when two mutually exclusive + flags have been supplied or when there are not enough non-flag + arguments. It is distinct from flags.Error which covers the lower + level of parsing and validating individual flags. + """ + + def __init__(self, message, exitcode=1): + super(UsageError, self).__init__(message) + self.exitcode = exitcode + + +class HelpFlag(flags.BooleanFlag): + """Special boolean flag that displays usage and raises SystemExit.""" + NAME = 'help' + SHORT_NAME = '?' + + def __init__(self): + super(HelpFlag, self).__init__( + self.NAME, False, 'show this help', + short_name=self.SHORT_NAME, allow_hide_cpp=True) + + def parse(self, arg): + if self._parse(arg): + usage(shorthelp=True, writeto_stdout=True) + # Advertise --helpfull on stdout, since usage() was on stdout. + print() + print('Try --helpfull to get a list of all flags.') + sys.exit(1) + + +class HelpshortFlag(HelpFlag): + """--helpshort is an alias for --help.""" + NAME = 'helpshort' + SHORT_NAME = None + + +class HelpfullFlag(flags.BooleanFlag): + """Display help for flags in the main module and all dependent modules.""" + + def __init__(self): + super(HelpfullFlag, self).__init__( + 'helpfull', False, 'show full help', allow_hide_cpp=True) + + def parse(self, arg): + if self._parse(arg): + usage(writeto_stdout=True) + sys.exit(1) + + +class HelpXMLFlag(flags.BooleanFlag): + """Similar to HelpfullFlag, but generates output in XML format.""" + + def __init__(self): + super(HelpXMLFlag, self).__init__( + 'helpxml', False, 'like --helpfull, but generates XML output', + allow_hide_cpp=True) + + def parse(self, arg): + if self._parse(arg): + flags.FLAGS.write_help_in_xml_format(sys.stdout) + sys.exit(1) + + +def parse_flags_with_usage(args): + """Tries to parse the flags, print usage, and exit if unparsable. + + Args: + args: [str], a non-empty list of the command line arguments including + program name. + + Returns: + [str], a non-empty list of remaining command line arguments after parsing + flags, including program name. + """ + try: + return FLAGS(args) + except flags.Error as error: + message = str(error) + if '\n' in message: + final_message = 'FATAL Flags parsing error:\n%s\n' % textwrap.indent( + message, ' ') + else: + final_message = 'FATAL Flags parsing error: %s\n' % message + sys.stderr.write(final_message) + sys.stderr.write('Pass --helpshort or --helpfull to see help on flags.\n') + sys.exit(1) + + +_define_help_flags_called = False + + +def define_help_flags(): + """Registers help flags. Idempotent.""" + # Use a global to ensure idempotence. + global _define_help_flags_called + + if not _define_help_flags_called: + flags.DEFINE_flag(HelpFlag()) + flags.DEFINE_flag(HelpshortFlag()) # alias for --help + flags.DEFINE_flag(HelpfullFlag()) + flags.DEFINE_flag(HelpXMLFlag()) + _define_help_flags_called = True + + +def _register_and_parse_flags_with_usage( + argv=None, + flags_parser=parse_flags_with_usage, +): + """Registers help flags, parses arguments and shows usage if appropriate. + + This also calls sys.exit(0) if flag --only_check_args is True. + + Args: + argv: [str], a non-empty list of the command line arguments including + program name, sys.argv is used if None. + flags_parser: Callable[[List[Text]], Any], the function used to parse flags. + The return value of this function is passed to `main` untouched. + It must guarantee FLAGS is parsed after this function is called. + + Returns: + The return value of `flags_parser`. When using the default `flags_parser`, + it returns the following: + [str], a non-empty list of remaining command line arguments after parsing + flags, including program name. + + Raises: + Error: Raised when flags_parser is called, but FLAGS is not parsed. + SystemError: Raised when it's called more than once. + """ + if _register_and_parse_flags_with_usage.done: + raise SystemError('Flag registration can be done only once.') + + define_help_flags() + + original_argv = sys.argv if argv is None else argv + args_to_main = flags_parser(original_argv) + if not FLAGS.is_parsed(): + raise Error('FLAGS must be parsed after flags_parser is called.') + + # Exit when told so. + if FLAGS.only_check_args: + sys.exit(0) + # Immediately after flags are parsed, bump verbosity to INFO if the flag has + # not been set. + if FLAGS['verbosity'].using_default_value: + FLAGS.verbosity = 0 + _register_and_parse_flags_with_usage.done = True + + return args_to_main + +_register_and_parse_flags_with_usage.done = False + + +def _run_main(main, argv): + """Calls main, optionally with pdb or profiler.""" + if FLAGS.run_with_pdb: + sys.exit(pdb.runcall(main, argv)) + elif FLAGS.run_with_profiling or FLAGS.profile_file: + # Avoid import overhead since most apps (including performance-sensitive + # ones) won't be run with profiling. + # pylint: disable=g-import-not-at-top + import atexit + if FLAGS.use_cprofile_for_profiling: + import cProfile as profile + else: + import profile + profiler = profile.Profile() + if FLAGS.profile_file: + atexit.register(profiler.dump_stats, FLAGS.profile_file) + else: + atexit.register(profiler.print_stats) + sys.exit(profiler.runcall(main, argv)) + else: + sys.exit(main(argv)) + + +def _call_exception_handlers(exception): + """Calls any installed exception handlers.""" + for handler in EXCEPTION_HANDLERS: + try: + if handler.wants(exception): + handler.handle(exception) + except: # pylint: disable=bare-except + try: + # We don't want to stop for exceptions in the exception handlers but + # we shouldn't hide them either. + logging.error(traceback.format_exc()) + except: # pylint: disable=bare-except + # In case even the logging statement fails, ignore. + pass + + +def run( + main, + argv=None, + flags_parser=parse_flags_with_usage, +): + """Begins executing the program. + + Args: + main: The main function to execute. It takes an single argument "argv", + which is a list of command line arguments with parsed flags removed. + The return value is passed to `sys.exit`, and so for example + a return value of 0 or None results in a successful termination, whereas + a return value of 1 results in abnormal termination. + For more details, see https://docs.python.org/3/library/sys#sys.exit + argv: A non-empty list of the command line arguments including program name, + sys.argv is used if None. + flags_parser: Callable[[List[Text]], Any], the function used to parse flags. + The return value of this function is passed to `main` untouched. + It must guarantee FLAGS is parsed after this function is called. + Should be passed as a keyword-only arg which will become mandatory in a + future release. + - Parses command line flags with the flag module. + - If there are any errors, prints usage(). + - Calls main() with the remaining arguments. + - If main() raises a UsageError, prints usage and the error message. + """ + try: + args = _run_init( + sys.argv if argv is None else argv, + flags_parser, + ) + while _init_callbacks: + callback = _init_callbacks.popleft() + callback() + try: + _run_main(main, args) + except UsageError as error: + usage(shorthelp=True, detailed_error=error, exitcode=error.exitcode) + except: + exc = sys.exc_info()[1] + # Don't try to post-mortem debug successful SystemExits, since those + # mean there wasn't actually an error. In particular, the test framework + # raises SystemExit(False) even if all tests passed. + if isinstance(exc, SystemExit) and not exc.code: + raise + + # Check the tty so that we don't hang waiting for input in an + # non-interactive scenario. + if FLAGS.pdb_post_mortem and sys.stdout.isatty(): + traceback.print_exc() + print() + print(' *** Entering post-mortem debugging ***') + print() + pdb.post_mortem() + raise + except Exception as e: + _call_exception_handlers(e) + raise + +# Callbacks which have been deferred until after _run_init has been called. +_init_callbacks = collections.deque() + + +def call_after_init(callback): + """Calls the given callback only once ABSL has finished initialization. + + If ABSL has already finished initialization when ``call_after_init`` is + called then the callback is executed immediately, otherwise `callback` is + stored to be executed after ``app.run`` has finished initializing (aka. just + before the main function is called). + + If called after ``app.run``, this is equivalent to calling ``callback()`` in + the caller thread. If called before ``app.run``, callbacks are run + sequentially (in an undefined order) in the same thread as ``app.run``. + + Args: + callback: a callable to be called once ABSL has finished initialization. + This may be immediate if initialization has already finished. It + takes no arguments and returns nothing. + """ + if _run_init.done: + callback() + else: + _init_callbacks.append(callback) + + +def _run_init( + argv, + flags_parser, +): + """Does one-time initialization and re-parses flags on rerun.""" + if _run_init.done: + return flags_parser(argv) + command_name.make_process_name_useful() + # Set up absl logging handler. + logging.use_absl_handler() + args = _register_and_parse_flags_with_usage( + argv=argv, + flags_parser=flags_parser, + ) + if faulthandler: + try: + faulthandler.enable() + except Exception: # pylint: disable=broad-except + # Some tests verify stderr output very closely, so don't print anything. + # Disabled faulthandler is a low-impact error. + pass + _run_init.done = True + return args + + +_run_init.done = False + + +def usage(shorthelp=False, writeto_stdout=False, detailed_error=None, + exitcode=None): + """Writes __main__'s docstring to stderr with some help text. + + Args: + shorthelp: bool, if True, prints only flags from the main module, + rather than all flags. + writeto_stdout: bool, if True, writes help message to stdout, + rather than to stderr. + detailed_error: str, additional detail about why usage info was presented. + exitcode: optional integer, if set, exits with this status code after + writing help. + """ + if writeto_stdout: + stdfile = sys.stdout + else: + stdfile = sys.stderr + + doc = sys.modules['__main__'].__doc__ + if not doc: + doc = '\nUSAGE: %s [flags]\n' % sys.argv[0] + doc = flags.text_wrap(doc, indent=' ', firstline_indent='') + else: + # Replace all '%s' with sys.argv[0], and all '%%' with '%'. + num_specifiers = doc.count('%') - 2 * doc.count('%%') + try: + doc %= (sys.argv[0],) * num_specifiers + except (OverflowError, TypeError, ValueError): + # Just display the docstring as-is. + pass + if shorthelp: + flag_str = FLAGS.main_module_help() + else: + flag_str = FLAGS.get_help() + try: + stdfile.write(doc) + if flag_str: + stdfile.write('\nflags:\n') + stdfile.write(flag_str) + stdfile.write('\n') + if detailed_error is not None: + stdfile.write('\n%s\n' % detailed_error) + except IOError as e: + # We avoid printing a huge backtrace if we get EPIPE, because + # "foo.par --help | less" is a frequent use case. + if e.errno != errno.EPIPE: + raise + if exitcode is not None: + sys.exit(exitcode) + + +class ExceptionHandler(object): + """Base exception handler from which other may inherit.""" + + def wants(self, exc): + """Returns whether this handler wants to handle the exception or not. + + This base class returns True for all exceptions by default. Override in + subclass if it wants to be more selective. + + Args: + exc: Exception, the current exception. + """ + del exc # Unused. + return True + + def handle(self, exc): + """Do something with the current exception. + + Args: + exc: Exception, the current exception + + This method must be overridden. + """ + raise NotImplementedError() + + +def install_exception_handler(handler): + """Installs an exception handler. + + Args: + handler: ExceptionHandler, the exception handler to install. + + Raises: + TypeError: Raised when the handler was not of the correct type. + + All installed exception handlers will be called if main() exits via + an abnormal exception, i.e. not one of SystemExit, KeyboardInterrupt, + FlagsError or UsageError. + """ + if not isinstance(handler, ExceptionHandler): + raise TypeError('handler of type %s does not inherit from ExceptionHandler' + % type(handler)) + EXCEPTION_HANDLERS.append(handler) diff --git a/venv/lib/python3.10/site-packages/absl/app.pyi b/venv/lib/python3.10/site-packages/absl/app.pyi new file mode 100644 index 0000000000000000000000000000000000000000..fe5e44809915f3dbd56b23207781a2219d86f842 --- /dev/null +++ b/venv/lib/python3.10/site-packages/absl/app.pyi @@ -0,0 +1,99 @@ + +from typing import Any, Callable, Collection, Iterable, List, NoReturn, Optional, Text, TypeVar, Union, overload + +from absl.flags import _flag + + +_MainArgs = TypeVar('_MainArgs') +_Exc = TypeVar('_Exc', bound=Exception) + + +class ExceptionHandler(): + + def wants(self, exc: _Exc) -> bool: + ... + + def handle(self, exc: _Exc): + ... + + +EXCEPTION_HANDLERS: List[ExceptionHandler] = ... + + +class HelpFlag(_flag.BooleanFlag): + def __init__(self): + ... + + +class HelpshortFlag(HelpFlag): + ... + + +class HelpfullFlag(_flag.BooleanFlag): + def __init__(self): + ... + + +class HelpXMLFlag(_flag.BooleanFlag): + def __init__(self): + ... + + +def define_help_flags() -> None: + ... + + +@overload +def usage(shorthelp: Union[bool, int] = ..., + writeto_stdout: Union[bool, int] = ..., + detailed_error: Optional[Any] = ..., + exitcode: None = ...) -> None: + ... + + +@overload +def usage(shorthelp: Union[bool, int] = ..., + writeto_stdout: Union[bool, int] = ..., + detailed_error: Optional[Any] = ..., + exitcode: int = ...) -> NoReturn: + ... + + +def install_exception_handler(handler: ExceptionHandler) -> None: + ... + + +class Error(Exception): + ... + + +class UsageError(Error): + exitcode: int + + +def parse_flags_with_usage(args: List[Text]) -> List[Text]: + ... + + +def call_after_init(callback: Callable[[], Any]) -> None: + ... + + +# Without the flag_parser argument, `main` should require a List[Text]. +@overload +def run( + main: Callable[[List[Text]], Any], + argv: Optional[List[Text]] = ..., + *, +) -> NoReturn: + ... + + +@overload +def run( + main: Callable[[_MainArgs], Any], + argv: Optional[List[Text]] = ..., + *, + flags_parser: Callable[[List[Text]], _MainArgs], +) -> NoReturn: + ... diff --git a/venv/lib/python3.10/site-packages/absl/command_name.py b/venv/lib/python3.10/site-packages/absl/command_name.py new file mode 100644 index 0000000000000000000000000000000000000000..9260fee9bd853ba33b2139b3d47b73e59c127f36 --- /dev/null +++ b/venv/lib/python3.10/site-packages/absl/command_name.py @@ -0,0 +1,63 @@ +# Copyright 2017 The Abseil Authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +"""A tiny stand alone library to change the kernel process name on Linux.""" + +import os +import sys + +# This library must be kept small and stand alone. It is used by small things +# that require no extension modules. + + +def make_process_name_useful(): + """Sets the process name to something better than 'python' if possible.""" + set_kernel_process_name(os.path.basename(sys.argv[0])) + + +def set_kernel_process_name(name): + """Changes the Kernel's /proc/self/status process name on Linux. + + The kernel name is NOT what will be shown by the ps or top command. + It is a 15 character string stored in the kernel's process table that + is included in the kernel log when a process is OOM killed. + The first 15 bytes of name are used. Non-ASCII unicode is replaced with '?'. + + Does nothing if /proc/self/comm cannot be written or prctl() fails. + + Args: + name: bytes|unicode, the Linux kernel's command name to set. + """ + if not isinstance(name, bytes): + name = name.encode('ascii', 'replace') + try: + # This is preferred to using ctypes to try and call prctl() when possible. + with open('/proc/self/comm', 'wb') as proc_comm: + proc_comm.write(name[:15]) + except EnvironmentError: + try: + import ctypes # pylint: disable=g-import-not-at-top + except ImportError: + return # No ctypes. + try: + libc = ctypes.CDLL('libc.so.6') + except EnvironmentError: + return # No libc.so.6. + pr_set_name = ctypes.c_ulong(15) # linux/prctl.h PR_SET_NAME value. + zero = ctypes.c_ulong(0) + try: + libc.prctl(pr_set_name, name, zero, zero, zero) + # Ignore the prctl return value. Nothing we can do if it errored. + except AttributeError: + return # No prctl. diff --git a/venv/lib/python3.10/site-packages/nvidia/cusparse/lib/libcusparse.so.12 b/venv/lib/python3.10/site-packages/nvidia/cusparse/lib/libcusparse.so.12 new file mode 100644 index 0000000000000000000000000000000000000000..f6c468c877574511b42458b4db54abb337ed64c0 --- /dev/null +++ b/venv/lib/python3.10/site-packages/nvidia/cusparse/lib/libcusparse.so.12 @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:500466a2f559de622a71bb920d3d2923e69135245747a2742ee13edff0ba6085 +size 264876688 diff --git a/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/INSTALLER b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/INSTALLER new file mode 100644 index 0000000000000000000000000000000000000000..a1b589e38a32041e49332e5e81c2d363dc418d68 --- /dev/null +++ b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/INSTALLER @@ -0,0 +1 @@ +pip diff --git a/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/License.txt b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/License.txt new file mode 100644 index 0000000000000000000000000000000000000000..b491c70e0aef319022ded661e111ddbd45b8a17f --- /dev/null +++ b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/License.txt @@ -0,0 +1,1568 @@ +End User License Agreement +-------------------------- + + +Preface +------- + +The Software License Agreement in Chapter 1 and the Supplement +in Chapter 2 contain license terms and conditions that govern +the use of NVIDIA software. By accepting this agreement, you +agree to comply with all the terms and conditions applicable +to the product(s) included herein. + + +NVIDIA Driver + + +Description + +This package contains the operating system driver and +fundamental system software components for NVIDIA GPUs. + + +NVIDIA CUDA Toolkit + + +Description + +The NVIDIA CUDA Toolkit provides command-line and graphical +tools for building, debugging and optimizing the performance +of applications accelerated by NVIDIA GPUs, runtime and math +libraries, and documentation including programming guides, +user manuals, and API references. + + +Default Install Location of CUDA Toolkit + +Windows platform: + +%ProgramFiles%\NVIDIA GPU Computing Toolkit\CUDA\v#.# + +Linux platform: + +/usr/local/cuda-#.# + +Mac platform: + +/Developer/NVIDIA/CUDA-#.# + + +NVIDIA CUDA Samples + + +Description + +This package includes over 100+ CUDA examples that demonstrate +various CUDA programming principles, and efficient CUDA +implementation of algorithms in specific application domains. + + +Default Install Location of CUDA Samples + +Windows platform: + +%ProgramData%\NVIDIA Corporation\CUDA Samples\v#.# + +Linux platform: + +/usr/local/cuda-#.#/samples + +and + +$HOME/NVIDIA_CUDA-#.#_Samples + +Mac platform: + +/Developer/NVIDIA/CUDA-#.#/samples + + +NVIDIA Nsight Visual Studio Edition (Windows only) + + +Description + +NVIDIA Nsight Development Platform, Visual Studio Edition is a +development environment integrated into Microsoft Visual +Studio that provides tools for debugging, profiling, analyzing +and optimizing your GPU computing and graphics applications. + + +Default Install Location of Nsight Visual Studio Edition + +Windows platform: + +%ProgramFiles(x86)%\NVIDIA Corporation\Nsight Visual Studio Edition #.# + + +1. License Agreement for NVIDIA Software Development Kits +--------------------------------------------------------- + + +Release Date: July 26, 2018 +--------------------------- + + +Important NoticeRead before downloading, installing, +copying or using the licensed software: +------------------------------------------------------- + +This license agreement, including exhibits attached +("Agreement”) is a legal agreement between you and NVIDIA +Corporation ("NVIDIA") and governs your use of a NVIDIA +software development kit (“SDK”). + +Each SDK has its own set of software and materials, but here +is a description of the types of items that may be included in +a SDK: source code, header files, APIs, data sets and assets +(examples include images, textures, models, scenes, videos, +native API input/output files), binary software, sample code, +libraries, utility programs, programming code and +documentation. + +This Agreement can be accepted only by an adult of legal age +of majority in the country in which the SDK is used. + +If you are entering into this Agreement on behalf of a company +or other legal entity, you represent that you have the legal +authority to bind the entity to this Agreement, in which case +“you” will mean the entity you represent. + +If you don’t have the required age or authority to accept +this Agreement, or if you don’t accept all the terms and +conditions of this Agreement, do not download, install or use +the SDK. + +You agree to use the SDK only for purposes that are permitted +by (a) this Agreement, and (b) any applicable law, regulation +or generally accepted practices or guidelines in the relevant +jurisdictions. + + +1.1. License + + +1.1.1. License Grant + +Subject to the terms of this Agreement, NVIDIA hereby grants +you a non-exclusive, non-transferable license, without the +right to sublicense (except as expressly provided in this +Agreement) to: + + 1. Install and use the SDK, + + 2. Modify and create derivative works of sample source code + delivered in the SDK, and + + 3. Distribute those portions of the SDK that are identified + in this Agreement as distributable, as incorporated in + object code format into a software application that meets + the distribution requirements indicated in this Agreement. + + +1.1.2. Distribution Requirements + +These are the distribution requirements for you to exercise +the distribution grant: + + 1. Your application must have material additional + functionality, beyond the included portions of the SDK. + + 2. The distributable portions of the SDK shall only be + accessed by your application. + + 3. The following notice shall be included in modifications + and derivative works of sample source code distributed: + “This software contains source code provided by NVIDIA + Corporation.” + + 4. Unless a developer tool is identified in this Agreement + as distributable, it is delivered for your internal use + only. + + 5. The terms under which you distribute your application + must be consistent with the terms of this Agreement, + including (without limitation) terms relating to the + license grant and license restrictions and protection of + NVIDIA’s intellectual property rights. Additionally, you + agree that you will protect the privacy, security and + legal rights of your application users. + + 6. You agree to notify NVIDIA in writing of any known or + suspected distribution or use of the SDK not in compliance + with the requirements of this Agreement, and to enforce + the terms of your agreements with respect to distributed + SDK. + + +1.1.3. Authorized Users + +You may allow employees and contractors of your entity or of +your subsidiary(ies) to access and use the SDK from your +secure network to perform work on your behalf. + +If you are an academic institution you may allow users +enrolled or employed by the academic institution to access and +use the SDK from your secure network. + +You are responsible for the compliance with the terms of this +Agreement by your authorized users. If you become aware that +your authorized users didn’t follow the terms of this +Agreement, you agree to take reasonable steps to resolve the +non-compliance and prevent new occurrences. + + +1.1.4. Pre-Release SDK + +The SDK versions identified as alpha, beta, preview or +otherwise as pre-release, may not be fully functional, may +contain errors or design flaws, and may have reduced or +different security, privacy, accessibility, availability, and +reliability standards relative to commercial versions of +NVIDIA software and materials. Use of a pre-release SDK may +result in unexpected results, loss of data, project delays or +other unpredictable damage or loss. + +You may use a pre-release SDK at your own risk, understanding +that pre-release SDKs are not intended for use in production +or business-critical systems. + +NVIDIA may choose not to make available a commercial version +of any pre-release SDK. NVIDIA may also choose to abandon +development and terminate the availability of a pre-release +SDK at any time without liability. + + +1.1.5. Updates + +NVIDIA may, at its option, make available patches, workarounds +or other updates to this SDK. Unless the updates are provided +with their separate governing terms, they are deemed part of +the SDK licensed to you as provided in this Agreement. You +agree that the form and content of the SDK that NVIDIA +provides may change without prior notice to you. While NVIDIA +generally maintains compatibility between versions, NVIDIA may +in some cases make changes that introduce incompatibilities in +future versions of the SDK. + + +1.1.6. Third Party Licenses + +The SDK may come bundled with, or otherwise include or be +distributed with, third party software licensed by a NVIDIA +supplier and/or open source software provided under an open +source license. Use of third party software is subject to the +third-party license terms, or in the absence of third party +terms, the terms of this Agreement. Copyright to third party +software is held by the copyright holders indicated in the +third-party software or license. + + +1.1.7. Reservation of Rights + +NVIDIA reserves all rights, title, and interest in and to the +SDK, not expressly granted to you under this Agreement. + + +1.2. Limitations + +The following license limitations apply to your use of the +SDK: + + 1. You may not reverse engineer, decompile or disassemble, + or remove copyright or other proprietary notices from any + portion of the SDK or copies of the SDK. + + 2. Except as expressly provided in this Agreement, you may + not copy, sell, rent, sublicense, transfer, distribute, + modify, or create derivative works of any portion of the + SDK. For clarity, you may not distribute or sublicense the + SDK as a stand-alone product. + + 3. Unless you have an agreement with NVIDIA for this + purpose, you may not indicate that an application created + with the SDK is sponsored or endorsed by NVIDIA. + + 4. You may not bypass, disable, or circumvent any + encryption, security, digital rights management or + authentication mechanism in the SDK. + + 5. You may not use the SDK in any manner that would cause it + to become subject to an open source software license. As + examples, licenses that require as a condition of use, + modification, and/or distribution that the SDK be: + + a. Disclosed or distributed in source code form; + + b. Licensed for the purpose of making derivative works; + or + + c. Redistributable at no charge. + + 6. Unless you have an agreement with NVIDIA for this + purpose, you may not use the SDK with any system or + application where the use or failure of the system or + application can reasonably be expected to threaten or + result in personal injury, death, or catastrophic loss. + Examples include use in avionics, navigation, military, + medical, life support or other life critical applications. + NVIDIA does not design, test or manufacture the SDK for + these critical uses and NVIDIA shall not be liable to you + or any third party, in whole or in part, for any claims or + damages arising from such uses. + + 7. You agree to defend, indemnify and hold harmless NVIDIA + and its affiliates, and their respective employees, + contractors, agents, officers and directors, from and + against any and all claims, damages, obligations, losses, + liabilities, costs or debt, fines, restitutions and + expenses (including but not limited to attorney’s fees + and costs incident to establishing the right of + indemnification) arising out of or related to your use of + the SDK outside of the scope of this Agreement, or not in + compliance with its terms. + + +1.3. Ownership + + 1. NVIDIA or its licensors hold all rights, title and + interest in and to the SDK and its modifications and + derivative works, including their respective intellectual + property rights, subject to your rights described in this + section. This SDK may include software and materials from + NVIDIA’s licensors, and these licensors are intended + third party beneficiaries that may enforce this Agreement + with respect to their intellectual property rights. + + 2. You hold all rights, title and interest in and to your + applications and your derivative works of the sample + source code delivered in the SDK, including their + respective intellectual property rights, subject to + NVIDIA’s rights described in this section. + + 3. You may, but don’t have to, provide to NVIDIA + suggestions, feature requests or other feedback regarding + the SDK, including possible enhancements or modifications + to the SDK. For any feedback that you voluntarily provide, + you hereby grant NVIDIA and its affiliates a perpetual, + non-exclusive, worldwide, irrevocable license to use, + reproduce, modify, license, sublicense (through multiple + tiers of sublicensees), and distribute (through multiple + tiers of distributors) it without the payment of any + royalties or fees to you. NVIDIA will use feedback at its + choice. NVIDIA is constantly looking for ways to improve + its products, so you may send feedback to NVIDIA through + the developer portal at https://developer.nvidia.com. + + +1.4. No Warranties + +THE SDK IS PROVIDED BY NVIDIA “AS IS” AND “WITH ALL +FAULTS.” TO THE MAXIMUM EXTENT PERMITTED BY LAW, NVIDIA AND +ITS AFFILIATES EXPRESSLY DISCLAIM ALL WARRANTIES OF ANY KIND +OR NATURE, WHETHER EXPRESS, IMPLIED OR STATUTORY, INCLUDING, +BUT NOT LIMITED TO, ANY WARRANTIES OF MERCHANTABILITY, FITNESS +FOR A PARTICULAR PURPOSE, TITLE, NON-INFRINGEMENT, OR THE +ABSENCE OF ANY DEFECTS THEREIN, WHETHER LATENT OR PATENT. NO +WARRANTY IS MADE ON THE BASIS OF TRADE USAGE, COURSE OF +DEALING OR COURSE OF TRADE. + + +1.5. Limitation of Liability + +TO THE MAXIMUM EXTENT PERMITTED BY LAW, NVIDIA AND ITS +AFFILIATES SHALL NOT BE LIABLE FOR ANY SPECIAL, INCIDENTAL, +PUNITIVE OR CONSEQUENTIAL DAMAGES, OR ANY LOST PROFITS, LOSS +OF USE, LOSS OF DATA OR LOSS OF GOODWILL, OR THE COSTS OF +PROCURING SUBSTITUTE PRODUCTS, ARISING OUT OF OR IN CONNECTION +WITH THIS AGREEMENT OR THE USE OR PERFORMANCE OF THE SDK, +WHETHER SUCH LIABILITY ARISES FROM ANY CLAIM BASED UPON BREACH +OF CONTRACT, BREACH OF WARRANTY, TORT (INCLUDING NEGLIGENCE), +PRODUCT LIABILITY OR ANY OTHER CAUSE OF ACTION OR THEORY OF +LIABILITY. IN NO EVENT WILL NVIDIA’S AND ITS AFFILIATES +TOTAL CUMULATIVE LIABILITY UNDER OR ARISING OUT OF THIS +AGREEMENT EXCEED US$10.00. THE NATURE OF THE LIABILITY OR THE +NUMBER OF CLAIMS OR SUITS SHALL NOT ENLARGE OR EXTEND THIS +LIMIT. + +These exclusions and limitations of liability shall apply +regardless if NVIDIA or its affiliates have been advised of +the possibility of such damages, and regardless of whether a +remedy fails its essential purpose. These exclusions and +limitations of liability form an essential basis of the +bargain between the parties, and, absent any of these +exclusions or limitations of liability, the provisions of this +Agreement, including, without limitation, the economic terms, +would be substantially different. + + +1.6. Termination + + 1. This Agreement will continue to apply until terminated by + either you or NVIDIA as described below. + + 2. If you want to terminate this Agreement, you may do so by + stopping to use the SDK. + + 3. NVIDIA may, at any time, terminate this Agreement if: + + a. (i) you fail to comply with any term of this + Agreement and the non-compliance is not fixed within + thirty (30) days following notice from NVIDIA (or + immediately if you violate NVIDIA’s intellectual + property rights); + + b. (ii) you commence or participate in any legal + proceeding against NVIDIA with respect to the SDK; or + + c. (iii) NVIDIA decides to no longer provide the SDK in + a country or, in NVIDIA’s sole discretion, the + continued use of it is no longer commercially viable. + + 4. Upon any termination of this Agreement, you agree to + promptly discontinue use of the SDK and destroy all copies + in your possession or control. Your prior distributions in + accordance with this Agreement are not affected by the + termination of this Agreement. Upon written request, you + will certify in writing that you have complied with your + commitments under this section. Upon any termination of + this Agreement all provisions survive except for the + license grant provisions. + + +1.7. General + +If you wish to assign this Agreement or your rights and +obligations, including by merger, consolidation, dissolution +or operation of law, contact NVIDIA to ask for permission. Any +attempted assignment not approved by NVIDIA in writing shall +be void and of no effect. NVIDIA may assign, delegate or +transfer this Agreement and its rights and obligations, and if +to a non-affiliate you will be notified. + +You agree to cooperate with NVIDIA and provide reasonably +requested information to verify your compliance with this +Agreement. + +This Agreement will be governed in all respects by the laws of +the United States and of the State of Delaware as those laws +are applied to contracts entered into and performed entirely +within Delaware by Delaware residents, without regard to the +conflicts of laws principles. The United Nations Convention on +Contracts for the International Sale of Goods is specifically +disclaimed. You agree to all terms of this Agreement in the +English language. + +The state or federal courts residing in Santa Clara County, +California shall have exclusive jurisdiction over any dispute +or claim arising out of this Agreement. Notwithstanding this, +you agree that NVIDIA shall still be allowed to apply for +injunctive remedies or an equivalent type of urgent legal +relief in any jurisdiction. + +If any court of competent jurisdiction determines that any +provision of this Agreement is illegal, invalid or +unenforceable, such provision will be construed as limited to +the extent necessary to be consistent with and fully +enforceable under the law and the remaining provisions will +remain in full force and effect. Unless otherwise specified, +remedies are cumulative. + +Each party acknowledges and agrees that the other is an +independent contractor in the performance of this Agreement. + +The SDK has been developed entirely at private expense and is +“commercial items” consisting of “commercial computer +software” and “commercial computer software +documentation” provided with RESTRICTED RIGHTS. Use, +duplication or disclosure by the U.S. Government or a U.S. +Government subcontractor is subject to the restrictions in +this Agreement pursuant to DFARS 227.7202-3(a) or as set forth +in subparagraphs (c)(1) and (2) of the Commercial Computer +Software - Restricted Rights clause at FAR 52.227-19, as +applicable. Contractor/manufacturer is NVIDIA, 2788 San Tomas +Expressway, Santa Clara, CA 95051. + +The SDK is subject to United States export laws and +regulations. You agree that you will not ship, transfer or +export the SDK into any country, or use the SDK in any manner, +prohibited by the United States Bureau of Industry and +Security or economic sanctions regulations administered by the +U.S. Department of Treasury’s Office of Foreign Assets +Control (OFAC), or any applicable export laws, restrictions or +regulations. These laws include restrictions on destinations, +end users and end use. By accepting this Agreement, you +confirm that you are not a resident or citizen of any country +currently embargoed by the U.S. and that you are not otherwise +prohibited from receiving the SDK. + +Any notice delivered by NVIDIA to you under this Agreement +will be delivered via mail, email or fax. You agree that any +notices that NVIDIA sends you electronically will satisfy any +legal communication requirements. Please direct your legal +notices or other correspondence to NVIDIA Corporation, 2788 +San Tomas Expressway, Santa Clara, California 95051, United +States of America, Attention: Legal Department. + +This Agreement and any exhibits incorporated into this +Agreement constitute the entire agreement of the parties with +respect to the subject matter of this Agreement and supersede +all prior negotiations or documentation exchanged between the +parties relating to this SDK license. Any additional and/or +conflicting terms on documents issued by you are null, void, +and invalid. Any amendment or waiver under this Agreement +shall be in writing and signed by representatives of both +parties. + + +2. CUDA Toolkit Supplement to Software License Agreement for +NVIDIA Software Development Kits +------------------------------------------------------------ + + +Release date: August 16, 2018 +----------------------------- + +The terms in this supplement govern your use of the NVIDIA +CUDA Toolkit SDK under the terms of your license agreement +(“Agreement”) as modified by this supplement. Capitalized +terms used but not defined below have the meaning assigned to +them in the Agreement. + +This supplement is an exhibit to the Agreement and is +incorporated as an integral part of the Agreement. In the +event of conflict between the terms in this supplement and the +terms in the Agreement, the terms in this supplement govern. + + +2.1. License Scope + +The SDK is licensed for you to develop applications only for +use in systems with NVIDIA GPUs. + + +2.2. Distribution + +The portions of the SDK that are distributable under the +Agreement are listed in Attachment A. + + +2.3. Operating Systems + +Those portions of the SDK designed exclusively for use on the +Linux or FreeBSD operating systems, or other operating systems +derived from the source code to these operating systems, may +be copied and redistributed for use in accordance with this +Agreement, provided that the object code files are not +modified in any way (except for unzipping of compressed +files). + + +2.4. Audio and Video Encoders and Decoders + +You acknowledge and agree that it is your sole responsibility +to obtain any additional third-party licenses required to +make, have made, use, have used, sell, import, and offer for +sale your products or services that include or incorporate any +third-party software and content relating to audio and/or +video encoders and decoders from, including but not limited +to, Microsoft, Thomson, Fraunhofer IIS, Sisvel S.p.A., +MPEG-LA, and Coding Technologies. NVIDIA does not grant to you +under this Agreement any necessary patent or other rights with +respect to any audio and/or video encoders and decoders. + + +2.5. Licensing + +If the distribution terms in this Agreement are not suitable +for your organization, or for any questions regarding this +Agreement, please contact NVIDIA at +nvidia-compute-license-questions@nvidia.com. + + +2.6. Attachment A + +The following portions of the SDK are distributable under the +Agreement: + +Component + +CUDA Runtime + +Windows + +cudart.dll, cudart_static.lib, cudadevrt.lib + +Mac OSX + +libcudart.dylib, libcudart_static.a, libcudadevrt.a + +Linux + +libcudart.so, libcudart_static.a, libcudadevrt.a + +Android + +libcudart.so, libcudart_static.a, libcudadevrt.a + +Component + +CUDA FFT Library + +Windows + +cufft.dll, cufftw.dll, cufft.lib, cufftw.lib + +Mac OSX + +libcufft.dylib, libcufft_static.a, libcufftw.dylib, +libcufftw_static.a + +Linux + +libcufft.so, libcufft_static.a, libcufftw.so, +libcufftw_static.a + +Android + +libcufft.so, libcufft_static.a, libcufftw.so, +libcufftw_static.a + +Component + +CUDA BLAS Library + +Windows + +cublas.dll, cublasLt.dll + +Mac OSX + +libcublas.dylib, libcublasLt.dylib, libcublas_static.a, +libcublasLt_static.a + +Linux + +libcublas.so, libcublasLt.so, libcublas_static.a, +libcublasLt_static.a + +Android + +libcublas.so, libcublasLt.so, libcublas_static.a, +libcublasLt_static.a + +Component + +NVIDIA "Drop-in" BLAS Library + +Windows + +nvblas.dll + +Mac OSX + +libnvblas.dylib + +Linux + +libnvblas.so + +Component + +CUDA Sparse Matrix Library + +Windows + +cusparse.dll, cusparse.lib + +Mac OSX + +libcusparse.dylib, libcusparse_static.a + +Linux + +libcusparse.so, libcusparse_static.a + +Android + +libcusparse.so, libcusparse_static.a + +Component + +CUDA Linear Solver Library + +Windows + +cusolver.dll, cusolver.lib + +Mac OSX + +libcusolver.dylib, libcusolver_static.a + +Linux + +libcusolver.so, libcusolver_static.a + +Android + +libcusolver.so, libcusolver_static.a + +Component + +CUDA Random Number Generation Library + +Windows + +curand.dll, curand.lib + +Mac OSX + +libcurand.dylib, libcurand_static.a + +Linux + +libcurand.so, libcurand_static.a + +Android + +libcurand.so, libcurand_static.a + +Component + +CUDA Accelerated Graph Library + +Component + +NVIDIA Performance Primitives Library + +Windows + +nppc.dll, nppc.lib, nppial.dll, nppial.lib, nppicc.dll, +nppicc.lib, nppicom.dll, nppicom.lib, nppidei.dll, +nppidei.lib, nppif.dll, nppif.lib, nppig.dll, nppig.lib, +nppim.dll, nppim.lib, nppist.dll, nppist.lib, nppisu.dll, +nppisu.lib, nppitc.dll, nppitc.lib, npps.dll, npps.lib + +Mac OSX + +libnppc.dylib, libnppc_static.a, libnppial.dylib, +libnppial_static.a, libnppicc.dylib, libnppicc_static.a, +libnppicom.dylib, libnppicom_static.a, libnppidei.dylib, +libnppidei_static.a, libnppif.dylib, libnppif_static.a, +libnppig.dylib, libnppig_static.a, libnppim.dylib, +libnppisu_static.a, libnppitc.dylib, libnppitc_static.a, +libnpps.dylib, libnpps_static.a + +Linux + +libnppc.so, libnppc_static.a, libnppial.so, +libnppial_static.a, libnppicc.so, libnppicc_static.a, +libnppicom.so, libnppicom_static.a, libnppidei.so, +libnppidei_static.a, libnppif.so, libnppif_static.a +libnppig.so, libnppig_static.a, libnppim.so, +libnppim_static.a, libnppist.so, libnppist_static.a, +libnppisu.so, libnppisu_static.a, libnppitc.so +libnppitc_static.a, libnpps.so, libnpps_static.a + +Android + +libnppc.so, libnppc_static.a, libnppial.so, +libnppial_static.a, libnppicc.so, libnppicc_static.a, +libnppicom.so, libnppicom_static.a, libnppidei.so, +libnppidei_static.a, libnppif.so, libnppif_static.a +libnppig.so, libnppig_static.a, libnppim.so, +libnppim_static.a, libnppist.so, libnppist_static.a, +libnppisu.so, libnppisu_static.a, libnppitc.so +libnppitc_static.a, libnpps.so, libnpps_static.a + +Component + +NVIDIA JPEG Library + +Linux + +libnvjpeg.so, libnvjpeg_static.a + +Component + +Internal common library required for statically linking to +cuBLAS, cuSPARSE, cuFFT, cuRAND, nvJPEG and NPP + +Mac OSX + +libculibos.a + +Linux + +libculibos.a + +Component + +NVIDIA Runtime Compilation Library and Header + +All + +nvrtc.h + +Windows + +nvrtc.dll, nvrtc-builtins.dll + +Mac OSX + +libnvrtc.dylib, libnvrtc-builtins.dylib + +Linux + +libnvrtc.so, libnvrtc-builtins.so + +Component + +NVIDIA Optimizing Compiler Library + +Windows + +nvvm.dll + +Mac OSX + +libnvvm.dylib + +Linux + +libnvvm.so + +Component + +NVIDIA Common Device Math Functions Library + +Windows + +libdevice.10.bc + +Mac OSX + +libdevice.10.bc + +Linux + +libdevice.10.bc + +Component + +CUDA Occupancy Calculation Header Library + +All + +cuda_occupancy.h + +Component + +CUDA Half Precision Headers + +All + +cuda_fp16.h, cuda_fp16.hpp + +Component + +CUDA Profiling Tools Interface (CUPTI) Library + +Windows + +cupti.dll + +Mac OSX + +libcupti.dylib + +Linux + +libcupti.so + +Component + +NVIDIA Tools Extension Library + +Windows + +nvToolsExt.dll, nvToolsExt.lib + +Mac OSX + +libnvToolsExt.dylib + +Linux + +libnvToolsExt.so + +Component + +NVIDIA CUDA Driver Libraries + +Linux + +libcuda.so, libnvidia-fatbinaryloader.so, +libnvidia-ptxjitcompiler.so + +The NVIDIA CUDA Driver Libraries are only distributable in +applications that meet this criteria: + + 1. The application was developed starting from a NVIDIA CUDA + container obtained from Docker Hub or the NVIDIA GPU + Cloud, and + + 2. The resulting application is packaged as a Docker + container and distributed to users on Docker Hub or the + NVIDIA GPU Cloud only. + + +2.7. Attachment B + + +Additional Licensing Obligations + +The following third party components included in the SOFTWARE +are licensed to Licensee pursuant to the following terms and +conditions: + + 1. Licensee's use of the GDB third party component is + subject to the terms and conditions of GNU GPL v3: + + This product includes copyrighted third-party software licensed + under the terms of the GNU General Public License v3 ("GPL v3"). + All third-party software packages are copyright by their respective + authors. GPL v3 terms and conditions are hereby incorporated into + the Agreement by this reference: http://www.gnu.org/licenses/gpl.txt + + Consistent with these licensing requirements, the software + listed below is provided under the terms of the specified + open source software licenses. To obtain source code for + software provided under licenses that require + redistribution of source code, including the GNU General + Public License (GPL) and GNU Lesser General Public License + (LGPL), contact oss-requests@nvidia.com. This offer is + valid for a period of three (3) years from the date of the + distribution of this product by NVIDIA CORPORATION. + + Component License + CUDA-GDB GPL v3 + + 2. Licensee represents and warrants that any and all third + party licensing and/or royalty payment obligations in + connection with Licensee's use of the H.264 video codecs + are solely the responsibility of Licensee. + + 3. Licensee's use of the Thrust library is subject to the + terms and conditions of the Apache License Version 2.0. + All third-party software packages are copyright by their + respective authors. Apache License Version 2.0 terms and + conditions are hereby incorporated into the Agreement by + this reference. + http://www.apache.org/licenses/LICENSE-2.0.html + + In addition, Licensee acknowledges the following notice: + Thrust includes source code from the Boost Iterator, + Tuple, System, and Random Number libraries. + + Boost Software License - Version 1.0 - August 17th, 2003 + . . . . + + Permission is hereby granted, free of charge, to any person or + organization obtaining a copy of the software and accompanying + documentation covered by this license (the "Software") to use, + reproduce, display, distribute, execute, and transmit the Software, + and to prepare derivative works of the Software, and to permit + third-parties to whom the Software is furnished to do so, all + subject to the following: + + The copyright notices in the Software and this entire statement, + including the above license grant, this restriction and the following + disclaimer, must be included in all copies of the Software, in whole + or in part, and all derivative works of the Software, unless such + copies or derivative works are solely in the form of machine-executable + object code generated by a source language processor. + + THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, + EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF + MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, TITLE AND + NON-INFRINGEMENT. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR + ANYONE DISTRIBUTING THE SOFTWARE BE LIABLE FOR ANY DAMAGES OR + OTHER LIABILITY, WHETHER IN CONTRACT, TORT OR OTHERWISE, ARISING + FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR + OTHER DEALINGS IN THE SOFTWARE. + + 4. Licensee's use of the LLVM third party component is + subject to the following terms and conditions: + + ====================================================== + LLVM Release License + ====================================================== + University of Illinois/NCSA + Open Source License + + Copyright (c) 2003-2010 University of Illinois at Urbana-Champaign. + All rights reserved. + + Developed by: + + LLVM Team + + University of Illinois at Urbana-Champaign + + http://llvm.org + + Permission is hereby granted, free of charge, to any person obtaining a copy + of this software and associated documentation files (the "Software"), to + deal with the Software without restriction, including without limitation the + rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + sell copies of the Software, and to permit persons to whom the Software is + furnished to do so, subject to the following conditions: + + * Redistributions of source code must retain the above copyright notice, + this list of conditions and the following disclaimers. + + * Redistributions in binary form must reproduce the above copyright + notice, this list of conditions and the following disclaimers in the + documentation and/or other materials provided with the distribution. + + * Neither the names of the LLVM Team, University of Illinois at Urbana- + Champaign, nor the names of its contributors may be used to endorse or + promote products derived from this Software without specific prior + written permission. + + THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL + THE CONTRIBUTORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR + OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, + ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER + DEALINGS WITH THE SOFTWARE. + + 5. Licensee's use (e.g. nvprof) of the PCRE third party + component is subject to the following terms and + conditions: + + ------------ + PCRE LICENCE + ------------ + PCRE is a library of functions to support regular expressions whose syntax + and semantics are as close as possible to those of the Perl 5 language. + Release 8 of PCRE is distributed under the terms of the "BSD" licence, as + specified below. The documentation for PCRE, supplied in the "doc" + directory, is distributed under the same terms as the software itself. The + basic library functions are written in C and are freestanding. Also + included in the distribution is a set of C++ wrapper functions, and a just- + in-time compiler that can be used to optimize pattern matching. These are + both optional features that can be omitted when the library is built. + + THE BASIC LIBRARY FUNCTIONS + --------------------------- + Written by: Philip Hazel + Email local part: ph10 + Email domain: cam.ac.uk + University of Cambridge Computing Service, + Cambridge, England. + Copyright (c) 1997-2012 University of Cambridge + All rights reserved. + + PCRE JUST-IN-TIME COMPILATION SUPPORT + ------------------------------------- + Written by: Zoltan Herczeg + Email local part: hzmester + Emain domain: freemail.hu + Copyright(c) 2010-2012 Zoltan Herczeg + All rights reserved. + + STACK-LESS JUST-IN-TIME COMPILER + -------------------------------- + Written by: Zoltan Herczeg + Email local part: hzmester + Emain domain: freemail.hu + Copyright(c) 2009-2012 Zoltan Herczeg + All rights reserved. + + THE C++ WRAPPER FUNCTIONS + ------------------------- + Contributed by: Google Inc. + Copyright (c) 2007-2012, Google Inc. + All rights reserved. + + THE "BSD" LICENCE + ----------------- + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions are met: + + * Redistributions of source code must retain the above copyright notice, + this list of conditions and the following disclaimer. + + * Redistributions in binary form must reproduce the above copyright + notice, this list of conditions and the following disclaimer in the + documentation and/or other materials provided with the distribution. + + * Neither the name of the University of Cambridge nor the name of Google + Inc. nor the names of their contributors may be used to endorse or + promote products derived from this software without specific prior + written permission. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" + AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE + ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE + LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS + INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN + CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) + ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + POSSIBILITY OF SUCH DAMAGE. + + 6. Some of the cuBLAS library routines were written by or + derived from code written by Vasily Volkov and are subject + to the Modified Berkeley Software Distribution License as + follows: + + Copyright (c) 2007-2009, Regents of the University of California + + All rights reserved. + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions are + met: + * Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above + copyright notice, this list of conditions and the following + disclaimer in the documentation and/or other materials provided + with the distribution. + * Neither the name of the University of California, Berkeley nor + the names of its contributors may be used to endorse or promote + products derived from this software without specific prior + written permission. + + THIS SOFTWARE IS PROVIDED BY THE AUTHOR "AS IS" AND ANY EXPRESS OR + IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, + INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) + HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, + STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING + IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + POSSIBILITY OF SUCH DAMAGE. + + 7. Some of the cuBLAS library routines were written by or + derived from code written by Davide Barbieri and are + subject to the Modified Berkeley Software Distribution + License as follows: + + Copyright (c) 2008-2009 Davide Barbieri @ University of Rome Tor Vergata. + + All rights reserved. + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions are + met: + * Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above + copyright notice, this list of conditions and the following + disclaimer in the documentation and/or other materials provided + with the distribution. + * The name of the author may not be used to endorse or promote + products derived from this software without specific prior + written permission. + + THIS SOFTWARE IS PROVIDED BY THE AUTHOR "AS IS" AND ANY EXPRESS OR + IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED + WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE + DISCLAIMED. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, + INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR + SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) + HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, + STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING + IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE + POSSIBILITY OF SUCH DAMAGE. + + 8. Some of the cuBLAS library routines were derived from + code developed by the University of Tennessee and are + subject to the Modified Berkeley Software Distribution + License as follows: + + Copyright (c) 2010 The University of Tennessee. + + All rights reserved. + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions are + met: + * Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above + copyright notice, this list of conditions and the following + disclaimer listed in this license in the documentation and/or + other materials provided with the distribution. + * Neither the name of the copyright holders nor the names of its + contributors may be used to endorse or promote products derived + from this software without specific prior written permission. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT + OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, + SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT + LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + + 9. Some of the cuBLAS library routines were written by or + derived from code written by Jonathan Hogg and are subject + to the Modified Berkeley Software Distribution License as + follows: + + Copyright (c) 2012, The Science and Technology Facilities Council (STFC). + + All rights reserved. + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions are + met: + * Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above + copyright notice, this list of conditions and the following + disclaimer in the documentation and/or other materials provided + with the distribution. + * Neither the name of the STFC nor the names of its contributors + may be used to endorse or promote products derived from this + software without specific prior written permission. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE STFC BE + LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR + CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF + SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR + BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, + WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE + OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN + IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + + 10. Some of the cuBLAS library routines were written by or + derived from code written by Ahmad M. Abdelfattah, David + Keyes, and Hatem Ltaief, and are subject to the Apache + License, Version 2.0, as follows: + + -- (C) Copyright 2013 King Abdullah University of Science and Technology + Authors: + Ahmad Abdelfattah (ahmad.ahmad@kaust.edu.sa) + David Keyes (david.keyes@kaust.edu.sa) + Hatem Ltaief (hatem.ltaief@kaust.edu.sa) + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions + are met: + + * Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above copyright + notice, this list of conditions and the following disclaimer in the + documentation and/or other materials provided with the distribution. + * Neither the name of the King Abdullah University of Science and + Technology nor the names of its contributors may be used to endorse + or promote products derived from this software without specific prior + written permission. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT + HOLDERS OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, + SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT + LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE + + 11. Some of the cuSPARSE library routines were written by or + derived from code written by Li-Wen Chang and are subject + to the NCSA Open Source License as follows: + + Copyright (c) 2012, University of Illinois. + + All rights reserved. + + Developed by: IMPACT Group, University of Illinois, http://impact.crhc.illinois.edu + + Permission is hereby granted, free of charge, to any person obtaining + a copy of this software and associated documentation files (the + "Software"), to deal with the Software without restriction, including + without limitation the rights to use, copy, modify, merge, publish, + distribute, sublicense, and/or sell copies of the Software, and to + permit persons to whom the Software is furnished to do so, subject to + the following conditions: + * Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above + copyright notice, this list of conditions and the following + disclaimers in the documentation and/or other materials provided + with the distribution. + * Neither the names of IMPACT Group, University of Illinois, nor + the names of its contributors may be used to endorse or promote + products derived from this Software without specific prior + written permission. + + THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, + EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF + MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND + NONINFRINGEMENT. IN NO EVENT SHALL THE CONTRIBUTORS OR COPYRIGHT + HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER + IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR + IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS WITH THE + SOFTWARE. + + 12. Some of the cuRAND library routines were written by or + derived from code written by Mutsuo Saito and Makoto + Matsumoto and are subject to the following license: + + Copyright (c) 2009, 2010 Mutsuo Saito, Makoto Matsumoto and Hiroshima + University. All rights reserved. + + Copyright (c) 2011 Mutsuo Saito, Makoto Matsumoto, Hiroshima + University and University of Tokyo. All rights reserved. + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions are + met: + * Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above + copyright notice, this list of conditions and the following + disclaimer in the documentation and/or other materials provided + with the distribution. + * Neither the name of the Hiroshima University nor the names of + its contributors may be used to endorse or promote products + derived from this software without specific prior written + permission. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT + OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, + SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT + LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + + 13. Some of the cuRAND library routines were derived from + code developed by D. E. Shaw Research and are subject to + the following license: + + Copyright 2010-2011, D. E. Shaw Research. + + All rights reserved. + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions are + met: + * Redistributions of source code must retain the above copyright + notice, this list of conditions, and the following disclaimer. + * Redistributions in binary form must reproduce the above + copyright notice, this list of conditions, and the following + disclaimer in the documentation and/or other materials provided + with the distribution. + * Neither the name of D. E. Shaw Research nor the names of its + contributors may be used to endorse or promote products derived + from this software without specific prior written permission. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT + OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, + SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT + LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + + 14. Some of the Math library routines were written by or + derived from code developed by Norbert Juffa and are + subject to the following license: + + Copyright (c) 2015-2017, Norbert Juffa + All rights reserved. + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions + are met: + + 1. Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + + 2. Redistributions in binary form must reproduce the above copyright + notice, this list of conditions and the following disclaimer in the + documentation and/or other materials provided with the distribution. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT + HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, + SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT + LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + + 15. Licensee's use of the lz4 third party component is + subject to the following terms and conditions: + + Copyright (C) 2011-2013, Yann Collet. + BSD 2-Clause License (http://www.opensource.org/licenses/bsd-license.php) + + Redistribution and use in source and binary forms, with or without + modification, are permitted provided that the following conditions are + met: + + * Redistributions of source code must retain the above copyright + notice, this list of conditions and the following disclaimer. + * Redistributions in binary form must reproduce the above + copyright notice, this list of conditions and the following disclaimer + in the documentation and/or other materials provided with the + distribution. + + THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS + "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT + LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR + A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT + OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, + SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT + LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, + DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY + THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + + 16. The NPP library uses code from the Boost Math Toolkit, + and is subject to the following license: + + Boost Software License - Version 1.0 - August 17th, 2003 + . . . . + + Permission is hereby granted, free of charge, to any person or + organization obtaining a copy of the software and accompanying + documentation covered by this license (the "Software") to use, + reproduce, display, distribute, execute, and transmit the Software, + and to prepare derivative works of the Software, and to permit + third-parties to whom the Software is furnished to do so, all + subject to the following: + + The copyright notices in the Software and this entire statement, + including the above license grant, this restriction and the following + disclaimer, must be included in all copies of the Software, in whole + or in part, and all derivative works of the Software, unless such + copies or derivative works are solely in the form of machine-executable + object code generated by a source language processor. + + THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, + EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF + MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE, TITLE AND + NON-INFRINGEMENT. IN NO EVENT SHALL THE COPYRIGHT HOLDERS OR + ANYONE DISTRIBUTING THE SOFTWARE BE LIABLE FOR ANY DAMAGES OR + OTHER LIABILITY, WHETHER IN CONTRACT, TORT OR OTHERWISE, ARISING + FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR + OTHER DEALINGS IN THE SOFTWARE. + + 17. Portions of the Nsight Eclipse Edition is subject to the + following license: + + The Eclipse Foundation makes available all content in this plug-in + ("Content"). Unless otherwise indicated below, the Content is provided + to you under the terms and conditions of the Eclipse Public License + Version 1.0 ("EPL"). A copy of the EPL is available at http:// + www.eclipse.org/legal/epl-v10.html. For purposes of the EPL, "Program" + will mean the Content. + + If you did not receive this Content directly from the Eclipse + Foundation, the Content is being redistributed by another party + ("Redistributor") and different terms and conditions may apply to your + use of any object code in the Content. Check the Redistributor's + license that was provided with the Content. If no such license exists, + contact the Redistributor. Unless otherwise indicated below, the terms + and conditions of the EPL still apply to any source code in the + Content and such source code may be obtained at http://www.eclipse.org. + + 18. Some of the cuBLAS library routines uses code from + OpenAI, which is subject to the following license: + + License URL + https://github.com/openai/openai-gemm/blob/master/LICENSE + + License Text + The MIT License + + Copyright (c) 2016 OpenAI (http://openai.com), 2016 Google Inc. + + Permission is hereby granted, free of charge, to any person obtaining a copy + of this software and associated documentation files (the "Software"), to deal + in the Software without restriction, including without limitation the rights + to use, copy, modify, merge, publish, distribute, sublicense, and/or sell + copies of the Software, and to permit persons to whom the Software is + furnished to do so, subject to the following conditions: + + The above copyright notice and this permission notice shall be included in + all copies or substantial portions of the Software. + + THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN + THE SOFTWARE. + + 19. Licensee's use of the Visual Studio Setup Configuration + Samples is subject to the following license: + + The MIT License (MIT) + Copyright (C) Microsoft Corporation. All rights reserved. + + Permission is hereby granted, free of charge, to any person + obtaining a copy of this software and associated documentation + files (the "Software"), to deal in the Software without restriction, + including without limitation the rights to use, copy, modify, merge, + publish, distribute, sublicense, and/or sell copies of the Software, + and to permit persons to whom the Software is furnished to do so, + subject to the following conditions: + + The above copyright notice and this permission notice shall be included + in all copies or substantial portions of the Software. + + THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS + OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. + + 20. Licensee's use of linmath.h header for CPU functions for + GL vector/matrix operations from lunarG is subject to the + Apache License Version 2.0. + + 21. The DX12-CUDA sample uses the d3dx12.h header, which is + subject to the MIT license . + +----------------- diff --git a/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/METADATA b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/METADATA new file mode 100644 index 0000000000000000000000000000000000000000..301ffe540f0242521617353143a2ee3ad715a44e --- /dev/null +++ b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/METADATA @@ -0,0 +1,35 @@ +Metadata-Version: 2.1 +Name: nvidia-curand-cu12 +Version: 10.3.2.106 +Summary: CURAND native runtime libraries +Home-page: https://developer.nvidia.com/cuda-zone +Author: Nvidia CUDA Installer Team +Author-email: cuda_installer@nvidia.com +License: NVIDIA Proprietary Software +Keywords: cuda,nvidia,runtime,machine learning,deep learning +Classifier: Development Status :: 4 - Beta +Classifier: Intended Audience :: Developers +Classifier: Intended Audience :: Education +Classifier: Intended Audience :: Science/Research +Classifier: License :: Other/Proprietary License +Classifier: Natural Language :: English +Classifier: Programming Language :: Python :: 3 +Classifier: Programming Language :: Python :: 3.5 +Classifier: Programming Language :: Python :: 3.6 +Classifier: Programming Language :: Python :: 3.7 +Classifier: Programming Language :: Python :: 3.8 +Classifier: Programming Language :: Python :: 3.9 +Classifier: Programming Language :: Python :: 3.10 +Classifier: Programming Language :: Python :: 3.11 +Classifier: Programming Language :: Python :: 3 :: Only +Classifier: Topic :: Scientific/Engineering +Classifier: Topic :: Scientific/Engineering :: Mathematics +Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence +Classifier: Topic :: Software Development +Classifier: Topic :: Software Development :: Libraries +Classifier: Operating System :: Microsoft :: Windows +Classifier: Operating System :: POSIX :: Linux +Requires-Python: >=3 +License-File: License.txt + +CURAND native runtime libraries diff --git a/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/RECORD b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/RECORD new file mode 100644 index 0000000000000000000000000000000000000000..43a7eb55b35a1eedd1510118923adb984cd455f2 --- /dev/null +++ b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/RECORD @@ -0,0 +1,32 @@ +nvidia/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 +nvidia/__pycache__/__init__.cpython-310.pyc,, +nvidia/curand/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 +nvidia/curand/__pycache__/__init__.cpython-310.pyc,, +nvidia/curand/include/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 +nvidia/curand/include/__pycache__/__init__.cpython-310.pyc,, +nvidia/curand/include/curand.h,sha256=y393HddG5_OdV-3cK2l2Q_fxzGbEc1cyCcYjkOE3oIk,43966 +nvidia/curand/include/curand_discrete.h,sha256=2qD3BkI622XEu0444wVP7HeYkKAx0Rjr2HDhqU4SA7E,3486 +nvidia/curand/include/curand_discrete2.h,sha256=ZrQTO5R9x83AMX88uq7M8M94DLSC5VEz0PAkfcwtQeg,10883 +nvidia/curand/include/curand_globals.h,sha256=bES1Kx0NrATXk1DReMMkqWrB062nOnaAp39y22wViXU,3717 +nvidia/curand/include/curand_kernel.h,sha256=SjfAeh13ybXIxiekcgczzua02kIAqETopJKRhYvCat8,53133 +nvidia/curand/include/curand_lognormal.h,sha256=-X-iNkJSzWpAYYjogm689EJTZfzore9sxU7ObddljLk,28142 +nvidia/curand/include/curand_mrg32k3a.h,sha256=ZVVREjGNsJQJ-3IzZZ_LKGtGteslicb8E0Aly49BKPs,170296 +nvidia/curand/include/curand_mtgp32.h,sha256=Qhrmx0pHWF-P2Uu5bKwYE9ymEWq3c7qBzCITVMaKMfI,7845 +nvidia/curand/include/curand_mtgp32_host.h,sha256=SXqzmSQkzTLSRJ4pojTg_TNCC3T-G89HdBK-boSDqr4,18274 +nvidia/curand/include/curand_mtgp32_kernel.h,sha256=ajZnXr5ZXnQExElf6LPpigrrKPTmMIZbRyTEnJ-BDhw,13731 +nvidia/curand/include/curand_mtgp32dc_p_11213.h,sha256=7_gGYUH47UugIAEt60vYH5nFa-QUwTpDwSEgLg9cZts,276889 +nvidia/curand/include/curand_normal.h,sha256=lnmYVk2fn0oEVWOytdKhXrHL36GLCjMnB8OnZeCaYcA,26953 +nvidia/curand/include/curand_normal_static.h,sha256=5K4iTC9AuSWCe1LVxuj_0y3BVjtp0bxO6hndv2rbmiw,4727 +nvidia/curand/include/curand_philox4x32_x.h,sha256=T21IP-Rdg3_tSVU9Je4dLKuwEqE4ovfwi7r1hOY92Dw,7166 +nvidia/curand/include/curand_poisson.h,sha256=KrhXOmO_D7aclnj8geIyHqdpSQwWHurS9V_pVtgzodM,25461 +nvidia/curand/include/curand_precalc.h,sha256=I6NZdgT42fMm9qSCtP-rlOAqt4Zsqgal0ajktcPmEak,1392393 +nvidia/curand/include/curand_uniform.h,sha256=gpmRgQu5r6ppgLTg60NXoDdVJS6wMUy6jC5bh8l04e8,17472 +nvidia/curand/lib/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0 +nvidia/curand/lib/__pycache__/__init__.cpython-310.pyc,, +nvidia/curand/lib/libcurand.so.10,sha256=Qah4DXdgqpHMpyMtCF9VGDx-bPrsq8LzLnGSYIMQNfw,96681392 +nvidia_curand_cu12-10.3.2.106.dist-info/INSTALLER,sha256=zuuue4knoyJ-UwPPXg8fezS7VCrXJQrAP7zeNuwvFQg,4 +nvidia_curand_cu12-10.3.2.106.dist-info/License.txt,sha256=rW9YU_ugyg0VnQ9Y1JrkmDDC-Mk_epJki5zpCttMbM0,59262 +nvidia_curand_cu12-10.3.2.106.dist-info/METADATA,sha256=TTSGosdrLNURQYusjHa9N5vwOD1zz1DnnZV0im74NT4,1507 +nvidia_curand_cu12-10.3.2.106.dist-info/RECORD,, +nvidia_curand_cu12-10.3.2.106.dist-info/WHEEL,sha256=-kQi_VMfvRQozZJT7HUPMfY-5vLo0LVTmAylNJ3Ft98,106 +nvidia_curand_cu12-10.3.2.106.dist-info/top_level.txt,sha256=fTkAtiFuL16nUrB9ytDDtpytz2t0B4NvYTnRzwAhO14,7 diff --git a/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/WHEEL b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/WHEEL new file mode 100644 index 0000000000000000000000000000000000000000..06e355fe0e3ed7077903f119ae6928a17da8eb6f --- /dev/null +++ b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/WHEEL @@ -0,0 +1,5 @@ +Wheel-Version: 1.0 +Generator: bdist_wheel (0.37.1) +Root-Is-Purelib: true +Tag: py3-none-manylinux1_x86_64 + diff --git a/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/top_level.txt b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/top_level.txt new file mode 100644 index 0000000000000000000000000000000000000000..862f7abf232cdfbb928609856247292e81c9decb --- /dev/null +++ b/venv/lib/python3.10/site-packages/nvidia_curand_cu12-10.3.2.106.dist-info/top_level.txt @@ -0,0 +1 @@ +nvidia diff --git a/venv/lib/python3.10/site-packages/pydantic_core/__pycache__/__init__.cpython-310.pyc b/venv/lib/python3.10/site-packages/pydantic_core/__pycache__/__init__.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..d719f89be3c4a3da20a698a4006a418997e9213f Binary files /dev/null and b/venv/lib/python3.10/site-packages/pydantic_core/__pycache__/__init__.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/pydantic_core/__pycache__/core_schema.cpython-310.pyc b/venv/lib/python3.10/site-packages/pydantic_core/__pycache__/core_schema.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..8393761bc13d60a9de418ca3fb41e3dd580680d8 Binary files /dev/null and b/venv/lib/python3.10/site-packages/pydantic_core/__pycache__/core_schema.cpython-310.pyc differ