diff --git a/.gitattributes b/.gitattributes index 2da7a7bb6dbccca54f3b72d8917ad01c810290de..2c53de45b592821cf2d43c1b1b117fdd4a3215d6 100644 --- a/.gitattributes +++ b/.gitattributes @@ -100,3 +100,8 @@ venv/lib/python3.10/site-packages/nvidia/cusolver/lib/libcusolverMg.so.11 filter venv/lib/python3.10/site-packages/pandas/_libs/join.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text venv/lib/python3.10/site-packages/nvidia/cufft/lib/libcufft.so.11 filter=lfs diff=lfs merge=lfs -text venv/lib/python3.10/site-packages/pandas/_libs/algos.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text +venv/lib/python3.10/site-packages/nvidia/cuda_nvrtc/lib/libnvrtc.so.12 filter=lfs diff=lfs merge=lfs -text +venv/lib/python3.10/site-packages/nvidia/cudnn/lib/libcudnn_adv_infer.so.8 filter=lfs diff=lfs merge=lfs -text +venv/lib/python3.10/site-packages/pandas/_libs/groupby.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text +venv/lib/python3.10/site-packages/pandas/_libs/interval.cpython-310-x86_64-linux-gnu.so filter=lfs diff=lfs merge=lfs -text +venv/lib/python3.10/site-packages/nvidia/cusolver/lib/libcusolver.so.11 filter=lfs diff=lfs merge=lfs -text diff --git a/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f1f3b70365b7e99bd6d8198237a7655553dd2a21 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715682597 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..01d243a1d918fc9835eec0f9c25174a941aea0d9 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log @@ -0,0 +1,28 @@ + +2024-05-14:10:29:58,590 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:10:30:03,656 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:10:30:03,660 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:10:30:03,660 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step100'} +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/core/register.py:145: UserWarning: "hpu:X" notation is not supported by Gaudi PyTorch intergration bridge. Please change to "hpu" without index (Triggered internally at /npu-stack/pytorch-integration/pytorch_helpers/lazy_to_backend.cpp:53.) + return func(*args, **kwargs) +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +[2024-05-14 10:30:12,782] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +2024-05-14:10:30:13,148 WARNING [task.py:763] [Task: indiccopa-hi] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-14:10:30:13,148 WARNING [task.py:775] [Task: indiccopa-hi] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for ai4bharat/IndicCOPA contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/ai4bharat/IndicCOPA +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-05-14:10:30:15,039 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:10:30:15,040 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:10:30:15,062 INFO [task.py:395] Building contexts for indiccopa-hi on rank 7... +100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 56/56 [00:00<00:00, 94519.53it/s] +Passed argument batch_size = auto:1. Detecting largest batch size +Determined largest batch size: 64 +2024-05-14:10:30:16,730 INFO [evaluator.py:379] Running loglikelihood requests +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..0b67b8703ff9fc9de2792bec5e5c55dae571c5fe --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T10:29:58.450520", + "startedAt": "2024-05-14T10:29:57.959514", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3395.2857171052633, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.001, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.005, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 76.92145538330078 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e6011967adf661d00f7d085970391223077a0c73 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 25}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e29f0def60d0ecc1e9e7215328c918c08e7c98a6 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/logs/debug-internal.log @@ -0,0 +1,193 @@ +2024-05-14 10:29:57,971 INFO StreamThr :1084 [internal.py:wandb_internal():85] W&B internal server running at pid: 1084, started at: 2024-05-14 10:29:57.971344 +2024-05-14 10:29:57,973 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: status +2024-05-14 10:29:57,974 INFO WriterThread:1084 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/run-agf96au8.wandb +2024-05-14 10:29:57,975 DEBUG SenderThread:1084 [sender.py:send():378] send: header +2024-05-14 10:29:57,985 DEBUG SenderThread:1084 [sender.py:send():378] send: run +2024-05-14 10:29:58,271 INFO SenderThread:1084 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files +2024-05-14 10:29:58,271 INFO SenderThread:1084 [sender.py:_start_run_threads():1123] run started: agf96au8 with start time 1715682597.970983 +2024-05-14 10:29:58,281 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 10:29:58,282 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: check_version +2024-05-14 10:29:58,379 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 10:29:58,381 DEBUG HandlerThread:1084 [system_info.py:__init__():26] System info init +2024-05-14 10:29:58,381 DEBUG HandlerThread:1084 [system_info.py:__init__():41] System info init done +2024-05-14 10:29:58,381 INFO HandlerThread:1084 [system_monitor.py:start():194] Starting system monitor +2024-05-14 10:29:58,381 INFO SystemMonitor:1084 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 10:29:58,382 INFO HandlerThread:1084 [system_monitor.py:probe():214] Collecting system info +2024-05-14 10:29:58,382 INFO SystemMonitor:1084 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 10:29:58,382 INFO SystemMonitor:1084 [interfaces.py:start():188] Started disk monitoring +2024-05-14 10:29:58,383 INFO SystemMonitor:1084 [interfaces.py:start():188] Started memory monitoring +2024-05-14 10:29:58,383 INFO SystemMonitor:1084 [interfaces.py:start():188] Started network monitoring +2024-05-14 10:29:58,450 DEBUG HandlerThread:1084 [system_info.py:probe():150] Probing system +2024-05-14 10:29:58,458 DEBUG HandlerThread:1084 [system_info.py:_probe_git():135] Probing git +2024-05-14 10:29:58,477 ERROR HandlerThread:1084 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 10:29:58,477 DEBUG HandlerThread:1084 [system_info.py:_probe_git():143] Probing git done +2024-05-14 10:29:58,477 DEBUG HandlerThread:1084 [system_info.py:probe():198] Probing system done +2024-05-14 10:29:58,477 DEBUG HandlerThread:1084 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T10:29:58.450520', 'startedAt': '2024-05-14T10:29:57.959514', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3395.2857171052633, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.001, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.005, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 76.92145538330078}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 10:29:58,477 INFO HandlerThread:1084 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 10:29:58,477 INFO HandlerThread:1084 [system_monitor.py:probe():227] Publishing system info +2024-05-14 10:29:58,479 INFO HandlerThread:1084 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 10:29:58,482 DEBUG SenderThread:1084 [sender.py:send():378] send: files +2024-05-14 10:29:58,482 INFO SenderThread:1084 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 10:29:58,582 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 10:29:58,582 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: python_packages +2024-05-14 10:29:58,584 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 10:29:58,584 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: stop_status +2024-05-14 10:29:58,779 DEBUG SenderThread:1084 [sender.py:send():378] send: telemetry +2024-05-14 10:29:59,059 INFO wandb-upload_0:1084 [upload_job.py:push():130] Uploaded file /tmp/tmpr_9r9vcjwandb/7l3hr8sf-wandb-metadata.json +2024-05-14 10:29:59,273 INFO Thread-12 :1084 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/requirements.txt +2024-05-14 10:29:59,273 INFO Thread-12 :1084 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-metadata.json +2024-05-14 10:29:59,273 INFO Thread-12 :1084 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:01,279 INFO Thread-12 :1084 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:03,658 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:05,283 INFO Thread-12 :1084 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:08,661 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:11,297 INFO Thread-12 :1084 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:13,583 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 10:30:13,583 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: stop_status +2024-05-14 10:30:13,706 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:15,304 INFO Thread-12 :1084 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:17,307 INFO Thread-12 :1084 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:19,231 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:19,309 INFO Thread-12 :1084 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:23,951 DEBUG SenderThread:1084 [sender.py:send():378] send: exit +2024-05-14 10:30:23,951 INFO SenderThread:1084 [sender.py:send_exit():585] handling exit code: 0 +2024-05-14 10:30:23,951 INFO SenderThread:1084 [sender.py:send_exit():587] handling runtime: 25 +2024-05-14 10:30:23,952 INFO SenderThread:1084 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:30:23,953 INFO SenderThread:1084 [sender.py:send_exit():593] send defer +2024-05-14 10:30:23,953 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,953 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 10:30:23,953 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,953 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 10:30:23,953 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 1 +2024-05-14 10:30:23,953 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,953 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 10:30:23,953 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,953 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 10:30:23,953 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 2 +2024-05-14 10:30:23,953 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,954 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 10:30:23,954 INFO HandlerThread:1084 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 10:30:23,954 DEBUG SystemMonitor:1084 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 10:30:23,954 INFO HandlerThread:1084 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 10:30:23,954 DEBUG SystemMonitor:1084 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 10:30:23,955 INFO HandlerThread:1084 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 10:30:23,955 DEBUG SystemMonitor:1084 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 10:30:23,955 INFO HandlerThread:1084 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 10:30:23,956 INFO HandlerThread:1084 [interfaces.py:finish():200] Joined network monitor +2024-05-14 10:30:23,957 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,957 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 10:30:23,957 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 3 +2024-05-14 10:30:23,957 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,957 DEBUG SenderThread:1084 [sender.py:send():378] send: stats +2024-05-14 10:30:23,957 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 10:30:23,958 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,958 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 10:30:23,958 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 4 +2024-05-14 10:30:23,958 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,958 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 10:30:23,958 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,958 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 10:30:23,958 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 5 +2024-05-14 10:30:23,958 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,958 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 10:30:23,958 DEBUG SenderThread:1084 [sender.py:send():378] send: summary +2024-05-14 10:30:23,959 INFO SenderThread:1084 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:30:23,959 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,959 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 10:30:23,959 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 6 +2024-05-14 10:30:23,959 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:23,959 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 10:30:23,959 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:23,959 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 10:30:23,962 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:24,138 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 7 +2024-05-14 10:30:24,138 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:24,138 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 10:30:24,138 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:24,138 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 10:30:24,313 INFO Thread-12 :1084 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/config.yaml +2024-05-14 10:30:24,314 INFO Thread-12 :1084 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-summary.json +2024-05-14 10:30:24,951 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:26,797 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 8 +2024-05-14 10:30:26,797 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:26,797 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:26,798 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 10:30:26,798 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:26,798 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 10:30:26,798 INFO SenderThread:1084 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 10:30:26,798 INFO SenderThread:1084 [job_builder.py:_get_source_type():576] no source found +2024-05-14 10:30:26,798 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 9 +2024-05-14 10:30:26,798 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:26,798 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 10:30:26,799 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:26,799 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 10:30:26,799 INFO SenderThread:1084 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 10:30:26,951 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:27,315 INFO SenderThread:1084 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:27,315 INFO SenderThread:1084 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files +2024-05-14 10:30:27,315 INFO SenderThread:1084 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log output.log +2024-05-14 10:30:27,315 INFO SenderThread:1084 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/requirements.txt requirements.txt +2024-05-14 10:30:27,317 INFO SenderThread:1084 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/config.yaml config.yaml +2024-05-14 10:30:27,317 INFO SenderThread:1084 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-metadata.json wandb-metadata.json +2024-05-14 10:30:27,318 INFO SenderThread:1084 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-summary.json wandb-summary.json +2024-05-14 10:30:27,319 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 10 +2024-05-14 10:30:27,322 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:27,322 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:27,322 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 10:30:27,322 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:27,322 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 10:30:27,322 INFO SenderThread:1084 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:30:27,553 INFO wandb-upload_1:1084 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/requirements.txt +2024-05-14 10:30:27,708 INFO wandb-upload_0:1084 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/output.log +2024-05-14 10:30:27,802 INFO wandb-upload_2:1084 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/config.yaml +2024-05-14 10:30:27,808 INFO wandb-upload_3:1084 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/files/wandb-summary.json +2024-05-14 10:30:27,952 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:27,952 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:28,008 INFO Thread-11 (_thread_body):1084 [sender.py:transition_state():613] send defer: 11 +2024-05-14 10:30:28,008 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:28,008 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 10:30:28,009 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:28,009 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 10:30:28,009 INFO SenderThread:1084 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 10:30:28,009 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 12 +2024-05-14 10:30:28,009 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:28,009 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 10:30:28,009 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:28,009 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 10:30:28,009 INFO SenderThread:1084 [file_stream.py:finish():601] file stream finish called +2024-05-14 10:30:28,094 INFO SenderThread:1084 [file_stream.py:finish():605] file stream finish is done +2024-05-14 10:30:28,094 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 13 +2024-05-14 10:30:28,094 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:28,094 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 10:30:28,094 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:28,094 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 10:30:28,094 INFO SenderThread:1084 [sender.py:transition_state():613] send defer: 14 +2024-05-14 10:30:28,095 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:28,095 DEBUG SenderThread:1084 [sender.py:send():378] send: final +2024-05-14 10:30:28,095 INFO HandlerThread:1084 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 10:30:28,095 DEBUG SenderThread:1084 [sender.py:send():378] send: footer +2024-05-14 10:30:28,095 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:28,095 INFO SenderThread:1084 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 10:30:28,095 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:28,095 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:28,096 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:30:28,096 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:30:28,096 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 10:30:28,096 DEBUG SenderThread:1084 [sender.py:send_request():405] send_request: server_info +2024-05-14 10:30:28,096 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 10:30:28,098 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 10:30:28,098 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 10:30:28,163 INFO MainThread:1084 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 10:30:28,163 INFO MainThread:1084 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 10:30:28,164 INFO MainThread:1084 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 10:30:28,164 DEBUG HandlerThread:1084 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 10:30:28,164 INFO HandlerThread:1084 [handler.py:finish():882] shutting down handler +2024-05-14 10:30:29,096 INFO WriterThread:1084 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/run-agf96au8.wandb +2024-05-14 10:30:29,163 INFO SenderThread:1084 [sender.py:finish():1545] shutting down sender +2024-05-14 10:30:29,163 INFO SenderThread:1084 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:30:29,163 INFO SenderThread:1084 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..5aca7f18915ef3a8455b55ae09ded86faec13a84 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 10:29:57,967 INFO MainThread:487 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 10:29:57,967 INFO MainThread:487 [wandb_setup.py:_flush():76] Configure stats pid to 487 +2024-05-14 10:29:57,967 INFO MainThread:487 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 10:29:57,967 INFO MainThread:487 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 10:29:57,968 WARNING MainThread:487 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/logs/debug.log +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/logs/debug-internal.log +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_init.py:init():560] calling init triggers +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_init.py:init():610] starting backend +2024-05-14 10:29:57,968 INFO MainThread:487 [wandb_init.py:init():614] setting up manager +2024-05-14 10:29:57,970 INFO MainThread:487 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 10:29:57,970 INFO MainThread:487 [wandb_init.py:init():622] backend started and connected +2024-05-14 10:29:57,973 INFO MainThread:487 [wandb_init.py:init():711] updated telemetry +2024-05-14 10:29:57,984 INFO MainThread:487 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 10:29:58,281 INFO MainThread:487 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 10:29:58,375 INFO MainThread:487 [wandb_run.py:_on_init():2405] got version response +2024-05-14 10:29:58,375 INFO MainThread:487 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 10:29:58,582 INFO MainThread:487 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 10:29:58,583 INFO MainThread:487 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 10:29:58,584 INFO MainThread:487 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 10:29:58,584 INFO MainThread:487 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 10:29:58,586 INFO MainThread:487 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 10:30:29,165 WARNING MsgRouterThr:487 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/run-agf96au8.wandb b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/run-agf96au8.wandb new file mode 100644 index 0000000000000000000000000000000000000000..b56ed83f1492dc2b93c821befe5c37a1719b831d Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_102957-agf96au8/run-agf96au8.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..a81e7c0eda0613364bf8ab01b84a5a6fcfbb9134 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715687084 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..b550b1d8001c532df6dd1178361767d09770d335 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log @@ -0,0 +1,28 @@ + +2024-05-14:11:44:44,813 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:11:44:49,803 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:11:44:49,805 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:11:44:49,806 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step100'} +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/core/register.py:145: UserWarning: "hpu:X" notation is not supported by Gaudi PyTorch intergration bridge. Please change to "hpu" without index (Triggered internally at /npu-stack/pytorch-integration/pytorch_helpers/lazy_to_backend.cpp:53.) + return func(*args, **kwargs) +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +2024-05-14:11:44:58,455 WARNING [task.py:763] [Task: indiccopa-hi] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-14:11:44:58,455 WARNING [task.py:775] [Task: indiccopa-hi] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +[2024-05-14 11:44:58,007] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for ai4bharat/IndicCOPA contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/ai4bharat/IndicCOPA +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-05-14:11:44:59,652 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:11:44:59,652 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:11:44:59,684 INFO [task.py:395] Building contexts for indiccopa-hi on rank 4... +100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 56/56 [00:00<00:00, 107842.53it/s] +Passed argument batch_size = auto:1. Detecting largest batch size +2024-05-14:11:45:05,430 INFO [evaluator.py:379] Running loglikelihood requests +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Determined largest batch size: 64 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..43b848cc09ee139f3f9bcddf2becfe984fe01f17 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T11:44:44.680232", + "startedAt": "2024-05-14T11:44:44.219830", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=global_step100" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3383.75652631579, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.996, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 77.77706527709961 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..f97bfbe43401ee851b2b689eb650f147f8bb41d3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 29}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..caff160e2fcf5d37bf23af79b3331fd307e4f8dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/logs/debug-internal.log @@ -0,0 +1,195 @@ +2024-05-14 11:44:44,231 INFO StreamThr :84148 [internal.py:wandb_internal():85] W&B internal server running at pid: 84148, started at: 2024-05-14 11:44:44.231029 +2024-05-14 11:44:44,233 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: status +2024-05-14 11:44:44,235 INFO WriterThread:84148 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/run-6mq0v4wa.wandb +2024-05-14 11:44:44,235 DEBUG SenderThread:84148 [sender.py:send():378] send: header +2024-05-14 11:44:44,246 DEBUG SenderThread:84148 [sender.py:send():378] send: run +2024-05-14 11:44:44,515 INFO SenderThread:84148 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files +2024-05-14 11:44:44,515 INFO SenderThread:84148 [sender.py:_start_run_threads():1123] run started: 6mq0v4wa with start time 1715687084.230895 +2024-05-14 11:44:44,522 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 11:44:44,523 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: check_version +2024-05-14 11:44:44,608 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 11:44:44,610 DEBUG HandlerThread:84148 [system_info.py:__init__():26] System info init +2024-05-14 11:44:44,610 DEBUG HandlerThread:84148 [system_info.py:__init__():41] System info init done +2024-05-14 11:44:44,610 INFO HandlerThread:84148 [system_monitor.py:start():194] Starting system monitor +2024-05-14 11:44:44,610 INFO SystemMonitor:84148 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 11:44:44,610 INFO HandlerThread:84148 [system_monitor.py:probe():214] Collecting system info +2024-05-14 11:44:44,611 INFO SystemMonitor:84148 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 11:44:44,611 INFO SystemMonitor:84148 [interfaces.py:start():188] Started disk monitoring +2024-05-14 11:44:44,612 INFO SystemMonitor:84148 [interfaces.py:start():188] Started memory monitoring +2024-05-14 11:44:44,612 INFO SystemMonitor:84148 [interfaces.py:start():188] Started network monitoring +2024-05-14 11:44:44,680 DEBUG HandlerThread:84148 [system_info.py:probe():150] Probing system +2024-05-14 11:44:44,687 DEBUG HandlerThread:84148 [system_info.py:_probe_git():135] Probing git +2024-05-14 11:44:44,707 ERROR HandlerThread:84148 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 11:44:44,707 DEBUG HandlerThread:84148 [system_info.py:_probe_git():143] Probing git done +2024-05-14 11:44:44,707 DEBUG HandlerThread:84148 [system_info.py:probe():198] Probing system done +2024-05-14 11:44:44,707 DEBUG HandlerThread:84148 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T11:44:44.680232', 'startedAt': '2024-05-14T11:44:44.219830', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=global_step100'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3383.75652631579, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.996, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 77.77706527709961}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 11:44:44,707 INFO HandlerThread:84148 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 11:44:44,707 INFO HandlerThread:84148 [system_monitor.py:probe():227] Publishing system info +2024-05-14 11:44:44,708 INFO HandlerThread:84148 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 11:44:44,712 DEBUG SenderThread:84148 [sender.py:send():378] send: files +2024-05-14 11:44:44,712 INFO SenderThread:84148 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 11:44:44,810 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 11:44:44,810 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 11:44:44,810 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: python_packages +2024-05-14 11:44:44,812 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: stop_status +2024-05-14 11:44:44,974 DEBUG SenderThread:84148 [sender.py:send():378] send: telemetry +2024-05-14 11:44:45,233 INFO wandb-upload_0:84148 [upload_job.py:push():130] Uploaded file /tmp/tmpay0_yn5gwandb/rpafhv4f-wandb-metadata.json +2024-05-14 11:44:45,516 INFO Thread-12 :84148 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-metadata.json +2024-05-14 11:44:45,516 INFO Thread-12 :84148 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/requirements.txt +2024-05-14 11:44:45,516 INFO Thread-12 :84148 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:44:47,522 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:44:49,805 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:44:51,527 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:44:54,807 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:44:59,546 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:44:59,810 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 11:44:59,811 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: stop_status +2024-05-14 11:44:59,987 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:45:00,558 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:45:01,558 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:45:04,987 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:45:06,576 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:45:07,577 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:45:08,600 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:45:09,989 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:45:14,525 DEBUG SenderThread:84148 [sender.py:send():378] send: exit +2024-05-14 11:45:14,525 INFO SenderThread:84148 [sender.py:send_exit():585] handling exit code: 0 +2024-05-14 11:45:14,525 INFO SenderThread:84148 [sender.py:send_exit():587] handling runtime: 29 +2024-05-14 11:45:14,527 INFO SenderThread:84148 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 11:45:14,527 INFO SenderThread:84148 [sender.py:send_exit():593] send defer +2024-05-14 11:45:14,527 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:14,528 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 11:45:14,528 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:14,528 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 11:45:14,528 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 1 +2024-05-14 11:45:14,528 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:14,528 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 11:45:14,528 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:14,528 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 11:45:14,528 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 2 +2024-05-14 11:45:14,528 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:14,528 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 11:45:14,528 INFO HandlerThread:84148 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 11:45:14,528 DEBUG SystemMonitor:84148 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 11:45:14,529 INFO HandlerThread:84148 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 11:45:14,529 DEBUG SystemMonitor:84148 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 11:45:14,529 INFO HandlerThread:84148 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 11:45:14,529 DEBUG SystemMonitor:84148 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 11:45:14,529 INFO HandlerThread:84148 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 11:45:14,531 INFO HandlerThread:84148 [interfaces.py:finish():200] Joined network monitor +2024-05-14 11:45:14,531 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:14,531 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 11:45:14,531 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 3 +2024-05-14 11:45:14,531 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:14,531 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 11:45:14,532 DEBUG SenderThread:84148 [sender.py:send():378] send: stats +2024-05-14 11:45:14,532 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:14,532 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 11:45:14,532 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 4 +2024-05-14 11:45:14,532 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:14,533 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 11:45:14,533 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:14,533 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 11:45:14,533 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 5 +2024-05-14 11:45:14,533 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:14,533 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 11:45:14,533 DEBUG SenderThread:84148 [sender.py:send():378] send: summary +2024-05-14 11:45:14,534 INFO SenderThread:84148 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 11:45:14,534 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:14,534 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 11:45:14,534 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 6 +2024-05-14 11:45:14,534 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:14,534 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 11:45:14,534 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:14,534 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 11:45:14,537 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:45:14,615 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/config.yaml +2024-05-14 11:45:14,616 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 7 +2024-05-14 11:45:14,616 INFO Thread-12 :84148 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-summary.json +2024-05-14 11:45:14,616 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:14,616 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 11:45:14,617 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:14,617 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 11:45:15,525 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 11:45:15,620 INFO Thread-12 :84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/config.yaml +2024-05-14 11:45:16,025 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 8 +2024-05-14 11:45:16,025 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 11:45:16,026 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:16,026 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 11:45:16,026 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:16,026 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 11:45:16,026 INFO SenderThread:84148 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 11:45:16,026 INFO SenderThread:84148 [job_builder.py:_get_source_type():576] no source found +2024-05-14 11:45:16,027 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 9 +2024-05-14 11:45:16,027 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:16,027 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 11:45:16,027 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:16,027 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 11:45:16,027 INFO SenderThread:84148 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 11:45:16,526 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 11:45:16,623 INFO SenderThread:84148 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:45:16,623 INFO SenderThread:84148 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files +2024-05-14 11:45:16,623 INFO SenderThread:84148 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-summary.json wandb-summary.json +2024-05-14 11:45:16,623 INFO SenderThread:84148 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/config.yaml config.yaml +2024-05-14 11:45:16,623 INFO SenderThread:84148 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-metadata.json wandb-metadata.json +2024-05-14 11:45:16,623 INFO SenderThread:84148 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log output.log +2024-05-14 11:45:16,625 INFO SenderThread:84148 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/requirements.txt requirements.txt +2024-05-14 11:45:16,625 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 10 +2024-05-14 11:45:16,625 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 11:45:16,625 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:16,626 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 11:45:16,629 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:16,629 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 11:45:16,629 INFO SenderThread:84148 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 11:45:16,863 INFO wandb-upload_0:84148 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/wandb-summary.json +2024-05-14 11:45:17,022 INFO wandb-upload_1:84148 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/config.yaml +2024-05-14 11:45:17,088 INFO wandb-upload_3:84148 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/requirements.txt +2024-05-14 11:45:17,130 INFO wandb-upload_2:84148 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/files/output.log +2024-05-14 11:45:17,330 INFO Thread-11 (_thread_body):84148 [sender.py:transition_state():613] send defer: 11 +2024-05-14 11:45:17,330 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:17,330 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 11:45:17,331 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:17,331 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 11:45:17,331 INFO SenderThread:84148 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 11:45:17,331 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 12 +2024-05-14 11:45:17,331 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:17,331 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 11:45:17,331 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:17,331 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 11:45:17,331 INFO SenderThread:84148 [file_stream.py:finish():601] file stream finish called +2024-05-14 11:45:17,406 INFO SenderThread:84148 [file_stream.py:finish():605] file stream finish is done +2024-05-14 11:45:17,406 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 13 +2024-05-14 11:45:17,407 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:17,407 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 11:45:17,407 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:17,407 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 11:45:17,407 INFO SenderThread:84148 [sender.py:transition_state():613] send defer: 14 +2024-05-14 11:45:17,407 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:45:17,407 INFO HandlerThread:84148 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 11:45:17,407 DEBUG SenderThread:84148 [sender.py:send():378] send: final +2024-05-14 11:45:17,407 DEBUG SenderThread:84148 [sender.py:send():378] send: footer +2024-05-14 11:45:17,407 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: defer +2024-05-14 11:45:17,407 INFO SenderThread:84148 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 11:45:17,408 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 11:45:17,408 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 11:45:17,408 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 11:45:17,409 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 11:45:17,409 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 11:45:17,409 DEBUG SenderThread:84148 [sender.py:send_request():405] send_request: server_info +2024-05-14 11:45:17,411 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 11:45:17,411 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 11:45:17,411 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 11:45:17,471 INFO MainThread:84148 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 11:45:17,471 INFO MainThread:84148 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 11:45:17,471 INFO MainThread:84148 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 11:45:17,471 DEBUG HandlerThread:84148 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 11:45:17,471 INFO HandlerThread:84148 [handler.py:finish():882] shutting down handler +2024-05-14 11:45:18,409 INFO WriterThread:84148 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/run-6mq0v4wa.wandb +2024-05-14 11:45:18,471 INFO SenderThread:84148 [sender.py:finish():1545] shutting down sender +2024-05-14 11:45:18,471 INFO SenderThread:84148 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 11:45:18,471 INFO SenderThread:84148 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..80bca3918d343923538087b090548c19f0af527b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 11:44:44,227 INFO MainThread:82829 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 11:44:44,227 INFO MainThread:82829 [wandb_setup.py:_flush():76] Configure stats pid to 82829 +2024-05-14 11:44:44,227 INFO MainThread:82829 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 11:44:44,227 INFO MainThread:82829 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 11:44:44,227 INFO MainThread:82829 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 11:44:44,227 INFO MainThread:82829 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 11:44:44,227 WARNING MainThread:82829 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 11:44:44,227 INFO MainThread:82829 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 11:44:44,227 INFO MainThread:82829 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 11:44:44,228 INFO MainThread:82829 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/logs/debug.log +2024-05-14 11:44:44,228 INFO MainThread:82829 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/logs/debug-internal.log +2024-05-14 11:44:44,228 INFO MainThread:82829 [wandb_init.py:init():560] calling init triggers +2024-05-14 11:44:44,228 INFO MainThread:82829 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 11:44:44,228 INFO MainThread:82829 [wandb_init.py:init():610] starting backend +2024-05-14 11:44:44,228 INFO MainThread:82829 [wandb_init.py:init():614] setting up manager +2024-05-14 11:44:44,229 INFO MainThread:82829 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 11:44:44,230 INFO MainThread:82829 [wandb_init.py:init():622] backend started and connected +2024-05-14 11:44:44,234 INFO MainThread:82829 [wandb_init.py:init():711] updated telemetry +2024-05-14 11:44:44,245 INFO MainThread:82829 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 11:44:44,522 INFO MainThread:82829 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 11:44:44,601 INFO MainThread:82829 [wandb_run.py:_on_init():2405] got version response +2024-05-14 11:44:44,601 INFO MainThread:82829 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 11:44:44,810 INFO MainThread:82829 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 11:44:44,810 INFO MainThread:82829 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 11:44:44,810 INFO MainThread:82829 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 11:44:44,810 INFO MainThread:82829 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 11:44:44,811 INFO MainThread:82829 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 11:45:18,472 WARNING MsgRouterThr:82829 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/run-6mq0v4wa.wandb b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/run-6mq0v4wa.wandb new file mode 100644 index 0000000000000000000000000000000000000000..59cca5d8bd82bd6892c003481734f73346194ecb Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_114444-6mq0v4wa/run-6mq0v4wa.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..46cf6f00b3bd7716effe94e46d60c979920d0178 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715705121 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..d17b4ac9eada603ced36d1c3877f7d77caa4f08f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log @@ -0,0 +1,33 @@ + +2024-05-14:16:45:22,412 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:16:45:27,005 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:16:45:27,006 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:16:45:27,007 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step120'} +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/data/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/data/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/data/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 928, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 631, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 686, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 369, in cached_file + raise EnvironmentError( +OSError: /data/cronscript/ckpts//hf_ckpt//global_step120 does not appear to have a file named config.json. Checkout 'https://huggingface.co//data/cronscript/ckpts//hf_ckpt//global_step120/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..d58820ae4d1d475f1f3533cfdcf836a85a6a7f05 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T16:45:22.278454", + "startedAt": "2024-05-14T16:45:21.801142", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3391.1009210526313, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3280.989, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3215.562, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 863.4300422668457 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 5}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..efb9d91775ebbcc72e5452ab6a05907045baa37a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/logs/debug-internal.log @@ -0,0 +1,182 @@ +2024-05-14 16:45:21,815 INFO StreamThr :137224 [internal.py:wandb_internal():85] W&B internal server running at pid: 137224, started at: 2024-05-14 16:45:21.814334 +2024-05-14 16:45:21,817 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: status +2024-05-14 16:45:21,819 INFO WriterThread:137224 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/run-09yzhcox.wandb +2024-05-14 16:45:21,819 DEBUG SenderThread:137224 [sender.py:send():378] send: header +2024-05-14 16:45:21,828 DEBUG SenderThread:137224 [sender.py:send():378] send: run +2024-05-14 16:45:22,093 INFO SenderThread:137224 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files +2024-05-14 16:45:22,093 INFO SenderThread:137224 [sender.py:_start_run_threads():1123] run started: 09yzhcox with start time 1715705121.813918 +2024-05-14 16:45:22,099 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 16:45:22,100 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: check_version +2024-05-14 16:45:22,188 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 16:45:22,190 DEBUG HandlerThread:137224 [system_info.py:__init__():26] System info init +2024-05-14 16:45:22,190 DEBUG HandlerThread:137224 [system_info.py:__init__():41] System info init done +2024-05-14 16:45:22,190 INFO HandlerThread:137224 [system_monitor.py:start():194] Starting system monitor +2024-05-14 16:45:22,190 INFO SystemMonitor:137224 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 16:45:22,190 INFO HandlerThread:137224 [system_monitor.py:probe():214] Collecting system info +2024-05-14 16:45:22,191 INFO SystemMonitor:137224 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 16:45:22,192 INFO SystemMonitor:137224 [interfaces.py:start():188] Started disk monitoring +2024-05-14 16:45:22,192 INFO SystemMonitor:137224 [interfaces.py:start():188] Started memory monitoring +2024-05-14 16:45:22,193 INFO SystemMonitor:137224 [interfaces.py:start():188] Started network monitoring +2024-05-14 16:45:22,278 DEBUG HandlerThread:137224 [system_info.py:probe():150] Probing system +2024-05-14 16:45:22,286 DEBUG HandlerThread:137224 [system_info.py:_probe_git():135] Probing git +2024-05-14 16:45:22,305 ERROR HandlerThread:137224 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 16:45:22,305 DEBUG HandlerThread:137224 [system_info.py:_probe_git():143] Probing git done +2024-05-14 16:45:22,305 DEBUG HandlerThread:137224 [system_info.py:probe():198] Probing system done +2024-05-14 16:45:22,305 DEBUG HandlerThread:137224 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T16:45:22.278454', 'startedAt': '2024-05-14T16:45:21.801142', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3391.1009210526313, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3280.989, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3215.562, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 863.4300422668457}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 16:45:22,306 INFO HandlerThread:137224 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 16:45:22,306 INFO HandlerThread:137224 [system_monitor.py:probe():227] Publishing system info +2024-05-14 16:45:22,307 INFO HandlerThread:137224 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 16:45:22,311 DEBUG SenderThread:137224 [sender.py:send():378] send: files +2024-05-14 16:45:22,311 INFO SenderThread:137224 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 16:45:22,408 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 16:45:22,408 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: python_packages +2024-05-14 16:45:22,409 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:45:22,409 DEBUG SenderThread:137224 [sender.py:send():378] send: telemetry +2024-05-14 16:45:22,410 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:45:22,832 INFO wandb-upload_0:137224 [upload_job.py:push():130] Uploaded file /tmp/tmpkoasite6wandb/kfzhka0i-wandb-metadata.json +2024-05-14 16:45:23,094 INFO Thread-12 :137224 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/requirements.txt +2024-05-14 16:45:23,094 INFO Thread-12 :137224 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log +2024-05-14 16:45:23,094 INFO Thread-12 :137224 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-metadata.json +2024-05-14 16:45:25,094 INFO Thread-12 :137224 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log +2024-05-14 16:45:27,006 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:45:28,167 DEBUG SenderThread:137224 [sender.py:send():378] send: exit +2024-05-14 16:45:28,167 INFO SenderThread:137224 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 16:45:28,168 INFO SenderThread:137224 [sender.py:send_exit():587] handling runtime: 5 +2024-05-14 16:45:28,168 INFO SenderThread:137224 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:45:28,169 INFO SenderThread:137224 [sender.py:send_exit():593] send defer +2024-05-14 16:45:28,169 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,169 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 16:45:28,169 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,169 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 16:45:28,169 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 1 +2024-05-14 16:45:28,169 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,169 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 16:45:28,169 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,169 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 16:45:28,169 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 2 +2024-05-14 16:45:28,169 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,169 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 16:45:28,170 INFO HandlerThread:137224 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 16:45:28,170 DEBUG SystemMonitor:137224 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 16:45:28,170 INFO HandlerThread:137224 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 16:45:28,170 DEBUG SystemMonitor:137224 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 16:45:28,170 INFO HandlerThread:137224 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 16:45:28,170 DEBUG SystemMonitor:137224 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 16:45:28,171 INFO HandlerThread:137224 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 16:45:28,172 INFO HandlerThread:137224 [interfaces.py:finish():200] Joined network monitor +2024-05-14 16:45:28,172 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,172 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 16:45:28,172 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 3 +2024-05-14 16:45:28,172 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,172 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 16:45:28,173 DEBUG SenderThread:137224 [sender.py:send():378] send: stats +2024-05-14 16:45:28,173 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,173 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 16:45:28,173 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 4 +2024-05-14 16:45:28,173 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,173 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 16:45:28,173 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,174 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 16:45:28,174 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 5 +2024-05-14 16:45:28,174 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,174 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 16:45:28,174 DEBUG SenderThread:137224 [sender.py:send():378] send: summary +2024-05-14 16:45:28,175 INFO SenderThread:137224 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:45:28,175 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,175 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 16:45:28,175 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 6 +2024-05-14 16:45:28,175 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,175 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 16:45:28,175 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,175 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 16:45:28,178 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:45:28,419 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 7 +2024-05-14 16:45:28,419 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,420 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 16:45:28,420 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,420 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 16:45:29,096 INFO Thread-12 :137224 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/config.yaml +2024-05-14 16:45:29,097 INFO Thread-12 :137224 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log +2024-05-14 16:45:29,097 INFO Thread-12 :137224 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-summary.json +2024-05-14 16:45:29,168 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:30,575 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 8 +2024-05-14 16:45:30,575 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:30,575 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:30,576 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 16:45:30,576 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:30,576 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 16:45:30,576 INFO SenderThread:137224 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 16:45:30,576 INFO SenderThread:137224 [job_builder.py:_get_source_type():576] no source found +2024-05-14 16:45:30,576 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 9 +2024-05-14 16:45:30,576 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:30,576 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 16:45:30,577 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:30,577 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 16:45:30,577 INFO SenderThread:137224 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 16:45:31,098 INFO SenderThread:137224 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log +2024-05-14 16:45:31,098 INFO SenderThread:137224 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files +2024-05-14 16:45:31,098 INFO SenderThread:137224 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-metadata.json wandb-metadata.json +2024-05-14 16:45:31,098 INFO SenderThread:137224 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log output.log +2024-05-14 16:45:31,098 INFO SenderThread:137224 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/requirements.txt requirements.txt +2024-05-14 16:45:31,098 INFO SenderThread:137224 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-summary.json wandb-summary.json +2024-05-14 16:45:31,099 INFO SenderThread:137224 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/config.yaml config.yaml +2024-05-14 16:45:31,099 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 10 +2024-05-14 16:45:31,102 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:31,102 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 16:45:31,104 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:31,105 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 16:45:31,105 INFO SenderThread:137224 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:45:31,168 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:31,168 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:31,324 INFO wandb-upload_1:137224 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/requirements.txt +2024-05-14 16:45:31,535 INFO wandb-upload_0:137224 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/output.log +2024-05-14 16:45:31,575 INFO wandb-upload_3:137224 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/config.yaml +2024-05-14 16:45:31,590 INFO wandb-upload_2:137224 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/files/wandb-summary.json +2024-05-14 16:45:31,790 INFO Thread-11 (_thread_body):137224 [sender.py:transition_state():613] send defer: 11 +2024-05-14 16:45:31,791 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:31,791 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 16:45:31,791 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:31,791 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 16:45:31,791 INFO SenderThread:137224 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 16:45:31,791 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 12 +2024-05-14 16:45:31,792 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:31,792 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 16:45:31,792 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:31,792 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 16:45:31,792 INFO SenderThread:137224 [file_stream.py:finish():601] file stream finish called +2024-05-14 16:45:32,024 INFO SenderThread:137224 [file_stream.py:finish():605] file stream finish is done +2024-05-14 16:45:32,024 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 13 +2024-05-14 16:45:32,024 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:32,024 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 16:45:32,024 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:32,025 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 16:45:32,025 INFO SenderThread:137224 [sender.py:transition_state():613] send defer: 14 +2024-05-14 16:45:32,025 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:32,025 INFO HandlerThread:137224 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 16:45:32,025 DEBUG SenderThread:137224 [sender.py:send():378] send: final +2024-05-14 16:45:32,025 DEBUG SenderThread:137224 [sender.py:send():378] send: footer +2024-05-14 16:45:32,025 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:32,025 INFO SenderThread:137224 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 16:45:32,026 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:32,026 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:32,026 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:32,026 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:32,027 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 16:45:32,027 DEBUG SenderThread:137224 [sender.py:send_request():405] send_request: server_info +2024-05-14 16:45:32,028 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 16:45:32,028 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 16:45:32,029 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 16:45:32,090 INFO MainThread:137224 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 16:45:32,090 INFO MainThread:137224 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 16:45:32,090 INFO MainThread:137224 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 16:45:32,090 DEBUG HandlerThread:137224 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 16:45:32,090 INFO HandlerThread:137224 [handler.py:finish():882] shutting down handler +2024-05-14 16:45:33,027 INFO WriterThread:137224 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/run-09yzhcox.wandb +2024-05-14 16:45:33,090 INFO SenderThread:137224 [sender.py:finish():1545] shutting down sender +2024-05-14 16:45:33,090 INFO SenderThread:137224 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:45:33,090 INFO SenderThread:137224 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a6be060c364cf99d21433123d050fb717b730390 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_setup.py:_flush():76] Configure stats pid to 136016 +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 16:45:21,811 WARNING MainThread:136016 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/logs/debug.log +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/logs/debug-internal.log +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_init.py:init():560] calling init triggers +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_init.py:init():610] starting backend +2024-05-14 16:45:21,811 INFO MainThread:136016 [wandb_init.py:init():614] setting up manager +2024-05-14 16:45:21,813 INFO MainThread:136016 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 16:45:21,813 INFO MainThread:136016 [wandb_init.py:init():622] backend started and connected +2024-05-14 16:45:21,816 INFO MainThread:136016 [wandb_init.py:init():711] updated telemetry +2024-05-14 16:45:21,827 INFO MainThread:136016 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 16:45:22,099 INFO MainThread:136016 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 16:45:22,178 INFO MainThread:136016 [wandb_run.py:_on_init():2405] got version response +2024-05-14 16:45:22,178 INFO MainThread:136016 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 16:45:22,408 INFO MainThread:136016 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 16:45:22,408 INFO MainThread:136016 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 16:45:22,409 INFO MainThread:136016 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 16:45:22,409 INFO MainThread:136016 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 16:45:22,410 INFO MainThread:136016 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 16:45:33,091 WARNING MsgRouterThr:136016 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/run-09yzhcox.wandb b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/run-09yzhcox.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ae6cd1ba5f063f195584e82c63dfb5cadb5f9177 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_164521-09yzhcox/run-09yzhcox.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d19354886ecb3e98878200b7598e9a0e035fc75f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716403396 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/output.log b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..b0dd9fea0076fc8ba4ede8e31f451180fd09de34 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/output.log @@ -0,0 +1,34 @@ + +2024-05-22:18:43:17,314 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:18:43:25,818 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:18:43:25,818 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:18:43:25,819 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000'} +2024-05-22:18:43:28,328 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..56f883ec74a071cf6c38965523373d342a6eadf7 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T18:43:17.073766", + "startedAt": "2024-05-22T18:43:16.472541", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step10000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.4363875000004, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.63845443725586 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..a6ad6556a5ae7c10b61198105c9f0ebfc6646de1 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 18:43:16,486 INFO MainThread:908 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 18:43:16,486 INFO MainThread:908 [wandb_setup.py:_flush():76] Configure stats pid to 908 +2024-05-22 18:43:16,486 INFO MainThread:908 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 18:43:16,486 INFO MainThread:908 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 18:43:16,486 INFO MainThread:908 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 18:43:16,486 INFO MainThread:908 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 18:43:16,486 WARNING MainThread:908 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 18:43:16,486 INFO MainThread:908 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 18:43:16,486 INFO MainThread:908 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 18:43:16,487 INFO MainThread:908 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/logs/debug.log +2024-05-22 18:43:16,487 INFO MainThread:908 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184316-86p21jxx/logs/debug-internal.log +2024-05-22 18:43:16,487 INFO MainThread:908 [wandb_init.py:init():560] calling init triggers +2024-05-22 18:43:16,487 INFO MainThread:908 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 18:43:16,487 INFO MainThread:908 [wandb_init.py:init():610] starting backend +2024-05-22 18:43:16,487 INFO MainThread:908 [wandb_init.py:init():614] setting up manager +2024-05-22 18:43:16,491 INFO MainThread:908 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 18:43:16,492 INFO MainThread:908 [wandb_init.py:init():622] backend started and connected +2024-05-22 18:43:16,496 INFO MainThread:908 [wandb_init.py:init():711] updated telemetry +2024-05-22 18:43:16,504 INFO MainThread:908 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 18:43:16,846 INFO MainThread:908 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 18:43:16,955 INFO MainThread:908 [wandb_run.py:_on_init():2405] got version response +2024-05-22 18:43:16,956 INFO MainThread:908 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 18:43:17,309 INFO MainThread:908 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 18:43:17,309 INFO MainThread:908 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 18:43:17,310 INFO MainThread:908 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 18:43:17,310 INFO MainThread:908 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 18:43:17,311 INFO MainThread:908 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 18:43:31,815 WARNING MsgRouterThr:908 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..34fed27684a4dbbea0fd4cb23d16623121418de9 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716403733 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..efd1ba95adda5ba477d89b415fff37739f0ef48f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log @@ -0,0 +1,34 @@ + +2024-05-22:18:48:53,708 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:18:49:02,380 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:18:49:02,381 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:18:49:02,381 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000'} +2024-05-22:18:49:04,670 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..9a56d1decd507e1058bee7d6a5a1fe11fb073f0a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T18:48:53.497474", + "startedAt": "2024-05-22T18:48:52.976415", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2326.78540625, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64408493041992 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..c0b8f64c843399aabbc135af30f6b1aadfd6d035 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-22 18:48:53,003 INFO StreamThr :1317 [internal.py:wandb_internal():85] W&B internal server running at pid: 1317, started at: 2024-05-22 18:48:53.001746 +2024-05-22 18:48:53,008 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: status +2024-05-22 18:48:53,010 INFO WriterThread:1317 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/run-tp70ku31.wandb +2024-05-22 18:48:53,011 DEBUG SenderThread:1317 [sender.py:send():378] send: header +2024-05-22 18:48:53,013 DEBUG SenderThread:1317 [sender.py:send():378] send: run +2024-05-22 18:48:53,268 INFO SenderThread:1317 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files +2024-05-22 18:48:53,269 INFO SenderThread:1317 [sender.py:_start_run_threads():1123] run started: tp70ku31 with start time 1716403733.001595 +2024-05-22 18:48:53,277 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 18:48:53,277 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: check_version +2024-05-22 18:48:53,400 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 18:48:53,402 DEBUG HandlerThread:1317 [system_info.py:__init__():26] System info init +2024-05-22 18:48:53,402 DEBUG HandlerThread:1317 [system_info.py:__init__():41] System info init done +2024-05-22 18:48:53,402 INFO HandlerThread:1317 [system_monitor.py:start():194] Starting system monitor +2024-05-22 18:48:53,403 INFO SystemMonitor:1317 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 18:48:53,403 INFO HandlerThread:1317 [system_monitor.py:probe():214] Collecting system info +2024-05-22 18:48:53,409 INFO SystemMonitor:1317 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 18:48:53,410 INFO SystemMonitor:1317 [interfaces.py:start():188] Started disk monitoring +2024-05-22 18:48:53,416 INFO SystemMonitor:1317 [interfaces.py:start():188] Started memory monitoring +2024-05-22 18:48:53,416 INFO SystemMonitor:1317 [interfaces.py:start():188] Started network monitoring +2024-05-22 18:48:53,497 DEBUG HandlerThread:1317 [system_info.py:probe():150] Probing system +2024-05-22 18:48:53,500 DEBUG HandlerThread:1317 [system_info.py:_probe_git():135] Probing git +2024-05-22 18:48:53,510 ERROR HandlerThread:1317 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 18:48:53,510 DEBUG HandlerThread:1317 [system_info.py:_probe_git():143] Probing git done +2024-05-22 18:48:53,510 DEBUG HandlerThread:1317 [system_info.py:probe():198] Probing system done +2024-05-22 18:48:53,510 DEBUG HandlerThread:1317 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T18:48:53.497474', 'startedAt': '2024-05-22T18:48:52.976415', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2326.78540625, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64408493041992}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 18:48:53,510 INFO HandlerThread:1317 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 18:48:53,510 INFO HandlerThread:1317 [system_monitor.py:probe():227] Publishing system info +2024-05-22 18:48:53,513 INFO HandlerThread:1317 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 18:48:53,519 DEBUG SenderThread:1317 [sender.py:send():378] send: files +2024-05-22 18:48:53,519 INFO SenderThread:1317 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 18:48:53,701 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 18:48:53,701 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: python_packages +2024-05-22 18:48:53,703 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 18:48:53,704 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: stop_status +2024-05-22 18:48:53,782 DEBUG SenderThread:1317 [sender.py:send():378] send: telemetry +2024-05-22 18:48:54,123 INFO wandb-upload_0:1317 [upload_job.py:push():130] Uploaded file /tmp/tmp03dd6wggwandb/wmis1f8f-wandb-metadata.json +2024-05-22 18:48:54,270 INFO Thread-12 :1317 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/requirements.txt +2024-05-22 18:48:54,270 INFO Thread-12 :1317 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log +2024-05-22 18:48:54,270 INFO Thread-12 :1317 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-metadata.json +2024-05-22 18:48:56,270 INFO Thread-12 :1317 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log +2024-05-22 18:48:58,794 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:49:04,277 INFO Thread-12 :1317 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log +2024-05-22 18:49:04,382 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:49:04,677 DEBUG SenderThread:1317 [sender.py:send():378] send: exit +2024-05-22 18:49:04,677 INFO SenderThread:1317 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 18:49:04,677 INFO SenderThread:1317 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 18:49:04,678 INFO SenderThread:1317 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:49:04,679 INFO SenderThread:1317 [sender.py:send_exit():593] send defer +2024-05-22 18:49:04,679 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:04,679 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 18:49:04,679 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:04,679 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 18:49:04,679 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 1 +2024-05-22 18:49:04,679 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:04,679 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 18:49:04,679 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:04,679 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 18:49:04,679 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 2 +2024-05-22 18:49:04,680 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:04,680 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 18:49:04,680 INFO HandlerThread:1317 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 18:49:04,680 DEBUG SystemMonitor:1317 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 18:49:04,680 DEBUG SystemMonitor:1317 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 18:49:04,681 DEBUG SystemMonitor:1317 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 18:49:04,683 INFO HandlerThread:1317 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 18:49:04,683 INFO HandlerThread:1317 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 18:49:04,683 INFO HandlerThread:1317 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 18:49:04,683 INFO HandlerThread:1317 [interfaces.py:finish():200] Joined network monitor +2024-05-22 18:49:04,684 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:04,684 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 18:49:04,684 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 3 +2024-05-22 18:49:04,684 DEBUG SenderThread:1317 [sender.py:send():378] send: stats +2024-05-22 18:49:04,685 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:04,685 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 18:49:04,685 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:04,685 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 18:49:04,685 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 4 +2024-05-22 18:49:04,685 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:04,685 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 18:49:04,686 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:04,686 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 18:49:04,686 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 5 +2024-05-22 18:49:04,686 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:04,686 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 18:49:04,686 DEBUG SenderThread:1317 [sender.py:send():378] send: summary +2024-05-22 18:49:04,687 INFO SenderThread:1317 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:49:04,687 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:04,687 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 18:49:04,687 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 6 +2024-05-22 18:49:04,687 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:04,687 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 18:49:04,687 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:04,687 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 18:49:04,692 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:49:04,778 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 7 +2024-05-22 18:49:04,778 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:04,778 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 18:49:04,778 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:04,778 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 18:49:05,279 INFO Thread-12 :1317 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/config.yaml +2024-05-22 18:49:05,280 INFO Thread-12 :1317 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-summary.json +2024-05-22 18:49:05,677 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:49:05,812 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 8 +2024-05-22 18:49:05,812 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:49:05,812 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:05,812 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 18:49:05,813 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:05,813 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 18:49:05,813 INFO SenderThread:1317 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 18:49:05,813 INFO SenderThread:1317 [job_builder.py:_get_source_type():576] no source found +2024-05-22 18:49:05,813 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 9 +2024-05-22 18:49:05,813 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:05,813 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 18:49:05,814 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:05,814 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 18:49:05,814 INFO SenderThread:1317 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 18:49:06,281 INFO SenderThread:1317 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log +2024-05-22 18:49:06,281 INFO SenderThread:1317 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files +2024-05-22 18:49:06,281 INFO SenderThread:1317 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/requirements.txt requirements.txt +2024-05-22 18:49:06,281 INFO SenderThread:1317 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-summary.json wandb-summary.json +2024-05-22 18:49:06,284 INFO SenderThread:1317 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log output.log +2024-05-22 18:49:06,284 INFO SenderThread:1317 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/config.yaml config.yaml +2024-05-22 18:49:06,284 INFO SenderThread:1317 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-metadata.json wandb-metadata.json +2024-05-22 18:49:06,284 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 10 +2024-05-22 18:49:06,284 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:06,284 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 18:49:06,286 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:06,286 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 18:49:06,286 INFO SenderThread:1317 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:49:06,522 INFO wandb-upload_0:1317 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/requirements.txt +2024-05-22 18:49:06,678 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:49:06,678 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:49:06,849 INFO wandb-upload_3:1317 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/config.yaml +2024-05-22 18:49:06,904 INFO wandb-upload_1:1317 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/wandb-summary.json +2024-05-22 18:49:06,912 INFO wandb-upload_2:1317 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/files/output.log +2024-05-22 18:49:07,112 INFO Thread-11 (_thread_body):1317 [sender.py:transition_state():613] send defer: 11 +2024-05-22 18:49:07,113 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:07,113 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 18:49:07,113 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:07,113 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 18:49:07,113 INFO SenderThread:1317 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 18:49:07,113 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 12 +2024-05-22 18:49:07,113 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:07,113 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 18:49:07,114 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:07,114 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 18:49:07,114 INFO SenderThread:1317 [file_stream.py:finish():601] file stream finish called +2024-05-22 18:49:07,188 INFO SenderThread:1317 [file_stream.py:finish():605] file stream finish is done +2024-05-22 18:49:07,188 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 13 +2024-05-22 18:49:07,188 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:07,188 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 18:49:07,189 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:07,189 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 18:49:07,189 INFO SenderThread:1317 [sender.py:transition_state():613] send defer: 14 +2024-05-22 18:49:07,189 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:07,189 INFO HandlerThread:1317 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 18:49:07,189 DEBUG SenderThread:1317 [sender.py:send():378] send: final +2024-05-22 18:49:07,189 DEBUG SenderThread:1317 [sender.py:send():378] send: footer +2024-05-22 18:49:07,189 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:07,189 INFO SenderThread:1317 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 18:49:07,190 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:49:07,190 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:49:07,190 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 18:49:07,190 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 18:49:07,190 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 18:49:07,191 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 18:49:07,191 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:49:07,191 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:49:07,191 DEBUG SenderThread:1317 [sender.py:send_request():405] send_request: server_info +2024-05-22 18:49:07,244 INFO MainThread:1317 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 18:49:07,244 INFO MainThread:1317 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 18:49:07,244 INFO MainThread:1317 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-22 18:49:07,245 DEBUG HandlerThread:1317 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 18:49:07,245 INFO HandlerThread:1317 [handler.py:finish():882] shutting down handler +2024-05-22 18:49:08,191 INFO WriterThread:1317 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/run-tp70ku31.wandb +2024-05-22 18:49:08,244 INFO SenderThread:1317 [sender.py:finish():1545] shutting down sender +2024-05-22 18:49:08,244 INFO SenderThread:1317 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:49:08,244 INFO SenderThread:1317 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..39b20ce56fd3d805515c2f20e303328d66c6f9cd --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 18:48:52,995 INFO MainThread:1162 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 18:48:52,995 INFO MainThread:1162 [wandb_setup.py:_flush():76] Configure stats pid to 1162 +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 18:48:52,996 WARNING MainThread:1162 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/logs/debug.log +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/logs/debug-internal.log +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_init.py:init():560] calling init triggers +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_init.py:init():610] starting backend +2024-05-22 18:48:52,996 INFO MainThread:1162 [wandb_init.py:init():614] setting up manager +2024-05-22 18:48:53,000 INFO MainThread:1162 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 18:48:53,001 INFO MainThread:1162 [wandb_init.py:init():622] backend started and connected +2024-05-22 18:48:53,005 INFO MainThread:1162 [wandb_init.py:init():711] updated telemetry +2024-05-22 18:48:53,013 INFO MainThread:1162 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 18:48:53,276 INFO MainThread:1162 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 18:48:53,394 INFO MainThread:1162 [wandb_run.py:_on_init():2405] got version response +2024-05-22 18:48:53,394 INFO MainThread:1162 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 18:48:53,702 INFO MainThread:1162 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 18:48:53,702 INFO MainThread:1162 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 18:48:53,703 INFO MainThread:1162 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 18:48:53,703 INFO MainThread:1162 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 18:48:53,705 INFO MainThread:1162 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 18:49:08,245 WARNING MsgRouterThr:1162 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/run-tp70ku31.wandb b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/run-tp70ku31.wandb new file mode 100644 index 0000000000000000000000000000000000000000..4f8d31d98827ea66adfb96b0523a47d489d2fb24 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_184852-tp70ku31/run-tp70ku31.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..ee145ae855faf8c50ad4d2336a69198ca8e3af2e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716404102 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..c634fd6fd90b5bc9c857c2d651a5f30a11e77a8b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log @@ -0,0 +1,34 @@ + +2024-05-22:18:55:03,262 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:18:55:11,773 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:18:55:11,775 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:18:55:11,775 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000'} +2024-05-22:18:55:14,076 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..35901a27b2b5b3b31b4851b439144d40f7f45265 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T18:55:03.061474", + "startedAt": "2024-05-22T18:55:02.549285", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2334.24860625, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3319.804, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3320.276, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3316.993, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3304.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.63605499267578 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d9e5ccd381b0ad7271628208e06292c7f63c0a6b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-22 18:55:02,573 INFO StreamThr :2076 [internal.py:wandb_internal():85] W&B internal server running at pid: 2076, started at: 2024-05-22 18:55:02.571012 +2024-05-22 18:55:02,577 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: status +2024-05-22 18:55:02,578 INFO WriterThread:2076 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/run-z3vhjirk.wandb +2024-05-22 18:55:02,580 DEBUG SenderThread:2076 [sender.py:send():378] send: header +2024-05-22 18:55:02,583 DEBUG SenderThread:2076 [sender.py:send():378] send: run +2024-05-22 18:55:02,848 INFO SenderThread:2076 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files +2024-05-22 18:55:02,848 INFO SenderThread:2076 [sender.py:_start_run_threads():1123] run started: z3vhjirk with start time 1716404102.571079 +2024-05-22 18:55:02,850 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 18:55:02,850 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: check_version +2024-05-22 18:55:02,965 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 18:55:02,968 DEBUG HandlerThread:2076 [system_info.py:__init__():26] System info init +2024-05-22 18:55:02,968 DEBUG HandlerThread:2076 [system_info.py:__init__():41] System info init done +2024-05-22 18:55:02,968 INFO HandlerThread:2076 [system_monitor.py:start():194] Starting system monitor +2024-05-22 18:55:02,968 INFO SystemMonitor:2076 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 18:55:02,968 INFO HandlerThread:2076 [system_monitor.py:probe():214] Collecting system info +2024-05-22 18:55:02,975 INFO SystemMonitor:2076 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 18:55:02,975 INFO SystemMonitor:2076 [interfaces.py:start():188] Started disk monitoring +2024-05-22 18:55:02,977 INFO SystemMonitor:2076 [interfaces.py:start():188] Started memory monitoring +2024-05-22 18:55:02,979 INFO SystemMonitor:2076 [interfaces.py:start():188] Started network monitoring +2024-05-22 18:55:03,061 DEBUG HandlerThread:2076 [system_info.py:probe():150] Probing system +2024-05-22 18:55:03,064 DEBUG HandlerThread:2076 [system_info.py:_probe_git():135] Probing git +2024-05-22 18:55:03,074 ERROR HandlerThread:2076 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 18:55:03,074 DEBUG HandlerThread:2076 [system_info.py:_probe_git():143] Probing git done +2024-05-22 18:55:03,074 DEBUG HandlerThread:2076 [system_info.py:probe():198] Probing system done +2024-05-22 18:55:03,074 DEBUG HandlerThread:2076 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T18:55:03.061474', 'startedAt': '2024-05-22T18:55:02.549285', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step18000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2334.24860625, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3319.804, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3320.276, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3316.993, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3304.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.63605499267578}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 18:55:03,075 INFO HandlerThread:2076 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 18:55:03,075 INFO HandlerThread:2076 [system_monitor.py:probe():227] Publishing system info +2024-05-22 18:55:03,078 INFO HandlerThread:2076 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 18:55:03,083 DEBUG SenderThread:2076 [sender.py:send():378] send: files +2024-05-22 18:55:03,083 INFO SenderThread:2076 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 18:55:03,256 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 18:55:03,256 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: python_packages +2024-05-22 18:55:03,259 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 18:55:03,259 DEBUG SenderThread:2076 [sender.py:send():378] send: telemetry +2024-05-22 18:55:03,260 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: stop_status +2024-05-22 18:55:03,665 INFO wandb-upload_0:2076 [upload_job.py:push():130] Uploaded file /tmp/tmph6kmpmp3wandb/08c0vywo-wandb-metadata.json +2024-05-22 18:55:03,850 INFO Thread-12 :2076 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/requirements.txt +2024-05-22 18:55:03,850 INFO Thread-12 :2076 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-metadata.json +2024-05-22 18:55:03,850 INFO Thread-12 :2076 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log +2024-05-22 18:55:05,849 INFO Thread-12 :2076 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log +2024-05-22 18:55:08,346 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:55:13,776 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:55:13,858 INFO Thread-12 :2076 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log +2024-05-22 18:55:14,083 DEBUG SenderThread:2076 [sender.py:send():378] send: exit +2024-05-22 18:55:14,084 INFO SenderThread:2076 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 18:55:14,084 INFO SenderThread:2076 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 18:55:14,085 INFO SenderThread:2076 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:55:14,085 INFO SenderThread:2076 [sender.py:send_exit():593] send defer +2024-05-22 18:55:14,085 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:14,085 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 18:55:14,086 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:14,086 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 18:55:14,086 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 1 +2024-05-22 18:55:14,086 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:14,086 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 18:55:14,086 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:14,086 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 18:55:14,086 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 2 +2024-05-22 18:55:14,086 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:14,086 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 18:55:14,086 INFO HandlerThread:2076 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 18:55:14,086 DEBUG SystemMonitor:2076 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 18:55:14,086 DEBUG SystemMonitor:2076 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 18:55:14,086 DEBUG SystemMonitor:2076 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 18:55:14,087 INFO HandlerThread:2076 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 18:55:14,087 INFO HandlerThread:2076 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 18:55:14,088 INFO HandlerThread:2076 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 18:55:14,088 INFO HandlerThread:2076 [interfaces.py:finish():200] Joined network monitor +2024-05-22 18:55:14,088 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:14,088 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 18:55:14,088 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 3 +2024-05-22 18:55:14,088 DEBUG SenderThread:2076 [sender.py:send():378] send: stats +2024-05-22 18:55:14,088 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:14,088 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 18:55:14,088 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:14,089 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 18:55:14,089 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 4 +2024-05-22 18:55:14,089 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:14,089 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 18:55:14,089 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:14,089 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 18:55:14,089 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 5 +2024-05-22 18:55:14,089 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:14,089 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 18:55:14,089 DEBUG SenderThread:2076 [sender.py:send():378] send: summary +2024-05-22 18:55:14,090 INFO SenderThread:2076 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:55:14,090 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:14,090 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 18:55:14,090 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 6 +2024-05-22 18:55:14,090 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:14,090 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 18:55:14,090 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:14,090 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 18:55:14,095 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:55:14,182 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 7 +2024-05-22 18:55:14,182 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:14,182 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 18:55:14,183 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:14,183 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 18:55:14,860 INFO Thread-12 :2076 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/config.yaml +2024-05-22 18:55:14,860 INFO Thread-12 :2076 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-summary.json +2024-05-22 18:55:15,083 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:55:15,370 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 8 +2024-05-22 18:55:15,370 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:55:15,370 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:15,370 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 18:55:15,370 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:15,370 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 18:55:15,370 INFO SenderThread:2076 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 18:55:15,371 INFO SenderThread:2076 [job_builder.py:_get_source_type():576] no source found +2024-05-22 18:55:15,371 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 9 +2024-05-22 18:55:15,371 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:15,371 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 18:55:15,371 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:15,371 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 18:55:15,371 INFO SenderThread:2076 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 18:55:15,861 INFO SenderThread:2076 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log +2024-05-22 18:55:15,861 INFO SenderThread:2076 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files +2024-05-22 18:55:15,861 INFO SenderThread:2076 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log output.log +2024-05-22 18:55:15,862 INFO SenderThread:2076 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/config.yaml config.yaml +2024-05-22 18:55:15,864 INFO SenderThread:2076 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-summary.json wandb-summary.json +2024-05-22 18:55:15,866 INFO SenderThread:2076 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/requirements.txt requirements.txt +2024-05-22 18:55:15,866 INFO SenderThread:2076 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-metadata.json wandb-metadata.json +2024-05-22 18:55:15,866 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 10 +2024-05-22 18:55:15,866 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:15,866 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 18:55:15,869 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:15,869 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 18:55:15,869 INFO SenderThread:2076 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:55:16,084 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:55:16,084 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:55:16,124 INFO wandb-upload_0:2076 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/output.log +2024-05-22 18:55:16,465 INFO wandb-upload_2:2076 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/wandb-summary.json +2024-05-22 18:55:16,480 INFO wandb-upload_1:2076 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/config.yaml +2024-05-22 18:55:16,486 INFO wandb-upload_3:2076 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/files/requirements.txt +2024-05-22 18:55:16,686 INFO Thread-11 (_thread_body):2076 [sender.py:transition_state():613] send defer: 11 +2024-05-22 18:55:16,686 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:16,686 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 18:55:16,686 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:16,686 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 18:55:16,687 INFO SenderThread:2076 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 18:55:16,687 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 12 +2024-05-22 18:55:16,687 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:16,687 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 18:55:16,687 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:16,687 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 18:55:16,687 INFO SenderThread:2076 [file_stream.py:finish():601] file stream finish called +2024-05-22 18:55:16,765 INFO SenderThread:2076 [file_stream.py:finish():605] file stream finish is done +2024-05-22 18:55:16,765 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 13 +2024-05-22 18:55:16,765 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:16,765 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 18:55:16,765 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:16,765 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 18:55:16,765 INFO SenderThread:2076 [sender.py:transition_state():613] send defer: 14 +2024-05-22 18:55:16,765 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:55:16,765 INFO HandlerThread:2076 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 18:55:16,765 DEBUG SenderThread:2076 [sender.py:send():378] send: final +2024-05-22 18:55:16,765 DEBUG SenderThread:2076 [sender.py:send():378] send: footer +2024-05-22 18:55:16,766 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: defer +2024-05-22 18:55:16,766 INFO SenderThread:2076 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 18:55:16,766 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:55:16,766 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:55:16,766 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 18:55:16,766 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 18:55:16,766 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 18:55:16,767 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 18:55:16,767 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:55:16,767 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:55:16,767 DEBUG SenderThread:2076 [sender.py:send_request():405] send_request: server_info +2024-05-22 18:55:16,822 INFO MainThread:2076 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 18:55:16,822 INFO MainThread:2076 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 18:55:16,822 INFO MainThread:2076 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-22 18:55:16,822 DEBUG HandlerThread:2076 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 18:55:16,822 INFO HandlerThread:2076 [handler.py:finish():882] shutting down handler +2024-05-22 18:55:17,767 INFO WriterThread:2076 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/run-z3vhjirk.wandb +2024-05-22 18:55:17,822 INFO SenderThread:2076 [sender.py:finish():1545] shutting down sender +2024-05-22 18:55:17,822 INFO SenderThread:2076 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:55:17,822 INFO SenderThread:2076 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..e3e922388668670ecda9eb990b3c756c12ba19fb --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_setup.py:_flush():76] Configure stats pid to 1921 +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 18:55:02,565 WARNING MainThread:1921 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/logs/debug.log +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/logs/debug-internal.log +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_init.py:init():560] calling init triggers +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_init.py:init():610] starting backend +2024-05-22 18:55:02,565 INFO MainThread:1921 [wandb_init.py:init():614] setting up manager +2024-05-22 18:55:02,570 INFO MainThread:1921 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 18:55:02,570 INFO MainThread:1921 [wandb_init.py:init():622] backend started and connected +2024-05-22 18:55:02,574 INFO MainThread:1921 [wandb_init.py:init():711] updated telemetry +2024-05-22 18:55:02,583 INFO MainThread:1921 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 18:55:02,849 INFO MainThread:1921 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 18:55:02,959 INFO MainThread:1921 [wandb_run.py:_on_init():2405] got version response +2024-05-22 18:55:02,959 INFO MainThread:1921 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 18:55:03,257 INFO MainThread:1921 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 18:55:03,257 INFO MainThread:1921 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 18:55:03,258 INFO MainThread:1921 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 18:55:03,258 INFO MainThread:1921 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 18:55:03,259 INFO MainThread:1921 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 18:55:17,823 WARNING MsgRouterThr:1921 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/run-z3vhjirk.wandb b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/run-z3vhjirk.wandb new file mode 100644 index 0000000000000000000000000000000000000000..9a75f763f7e47471dfd3efd7ae8fa7f716c11c1e Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_185502-z3vhjirk/run-z3vhjirk.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..80abd5a4077070b725f5f73916e2bfad38de12bf --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716436977 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..f3b5bc998d91812291c85aac36c8b2da09ec9f90 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:04:02:58,205 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:04:03:07,643 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:04:03:07,644 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:04:03:07,644 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100'} +2024-05-23:04:03:09,965 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..88181d04cb90f3bd8f00a85cc517ce4f45bd5aed --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/requirements.txt @@ -0,0 +1,156 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +megatron-lm==1.1.5 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6afbcf0de758f8f9d2d0647482ce67233ecfce30 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T04:02:57.991287", + "startedAt": "2024-05-23T04:02:57.450179", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.47999375, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3342.291, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.66277694702148 + } + }, + "memory": { + "total": 1007.4379577636719 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..018b4068e81a342faa2a3b0691dee6965106bcc3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 12}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..0c4811bf0432aa60c12d1bebf91dab80bb619cf6 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 04:02:57,471 INFO StreamThr :811 [internal.py:wandb_internal():85] W&B internal server running at pid: 811, started at: 2024-05-23 04:02:57.470037 +2024-05-23 04:02:57,476 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: status +2024-05-23 04:02:57,476 INFO WriterThread:811 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/run-0kj3nw8j.wandb +2024-05-23 04:02:57,479 DEBUG SenderThread:811 [sender.py:send():378] send: header +2024-05-23 04:02:57,483 DEBUG SenderThread:811 [sender.py:send():378] send: run +2024-05-23 04:02:57,792 INFO SenderThread:811 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files +2024-05-23 04:02:57,792 INFO SenderThread:811 [sender.py:_start_run_threads():1123] run started: 0kj3nw8j with start time 1716436977.469893 +2024-05-23 04:02:57,795 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 04:02:57,795 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: check_version +2024-05-23 04:02:57,915 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 04:02:57,918 DEBUG HandlerThread:811 [system_info.py:__init__():26] System info init +2024-05-23 04:02:57,918 DEBUG HandlerThread:811 [system_info.py:__init__():41] System info init done +2024-05-23 04:02:57,918 INFO HandlerThread:811 [system_monitor.py:start():194] Starting system monitor +2024-05-23 04:02:57,918 INFO SystemMonitor:811 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 04:02:57,918 INFO HandlerThread:811 [system_monitor.py:probe():214] Collecting system info +2024-05-23 04:02:57,925 INFO SystemMonitor:811 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 04:02:57,925 INFO SystemMonitor:811 [interfaces.py:start():188] Started disk monitoring +2024-05-23 04:02:57,926 INFO SystemMonitor:811 [interfaces.py:start():188] Started memory monitoring +2024-05-23 04:02:57,928 INFO SystemMonitor:811 [interfaces.py:start():188] Started network monitoring +2024-05-23 04:02:57,991 DEBUG HandlerThread:811 [system_info.py:probe():150] Probing system +2024-05-23 04:02:57,994 DEBUG HandlerThread:811 [system_info.py:_probe_git():135] Probing git +2024-05-23 04:02:58,004 ERROR HandlerThread:811 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 04:02:58,004 DEBUG HandlerThread:811 [system_info.py:_probe_git():143] Probing git done +2024-05-23 04:02:58,004 DEBUG HandlerThread:811 [system_info.py:probe():198] Probing system done +2024-05-23 04:02:58,004 DEBUG HandlerThread:811 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T04:02:57.991287', 'startedAt': '2024-05-23T04:02:57.450179', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.47999375, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3342.291, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.66277694702148}}, 'memory': {'total': 1007.4379577636719}} +2024-05-23 04:02:58,004 INFO HandlerThread:811 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 04:02:58,004 INFO HandlerThread:811 [system_monitor.py:probe():227] Publishing system info +2024-05-23 04:02:58,008 INFO HandlerThread:811 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 04:02:58,013 DEBUG SenderThread:811 [sender.py:send():378] send: files +2024-05-23 04:02:58,013 INFO SenderThread:811 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 04:02:58,197 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 04:02:58,197 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: python_packages +2024-05-23 04:02:58,198 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 04:02:58,200 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: stop_status +2024-05-23 04:02:58,361 DEBUG SenderThread:811 [sender.py:send():378] send: telemetry +2024-05-23 04:02:58,592 INFO wandb-upload_0:811 [upload_job.py:push():130] Uploaded file /tmp/tmp1bfxq6xawandb/trgvpg8r-wandb-metadata.json +2024-05-23 04:02:58,794 INFO Thread-12 :811 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log +2024-05-23 04:02:58,794 INFO Thread-12 :811 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/requirements.txt +2024-05-23 04:02:58,795 INFO Thread-12 :811 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-metadata.json +2024-05-23 04:03:00,794 INFO Thread-12 :811 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log +2024-05-23 04:03:03,363 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 04:03:08,645 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 04:03:08,804 INFO Thread-12 :811 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log +2024-05-23 04:03:09,975 DEBUG SenderThread:811 [sender.py:send():378] send: exit +2024-05-23 04:03:09,975 INFO SenderThread:811 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 04:03:09,975 INFO SenderThread:811 [sender.py:send_exit():587] handling runtime: 12 +2024-05-23 04:03:09,977 INFO SenderThread:811 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 04:03:09,977 INFO SenderThread:811 [sender.py:send_exit():593] send defer +2024-05-23 04:03:09,978 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:09,978 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 04:03:09,978 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:09,978 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 04:03:09,978 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 1 +2024-05-23 04:03:09,978 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:09,978 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 04:03:09,978 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:09,978 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 04:03:09,978 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 2 +2024-05-23 04:03:09,978 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:09,978 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 04:03:09,978 INFO HandlerThread:811 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 04:03:09,978 DEBUG SystemMonitor:811 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 04:03:09,979 DEBUG SystemMonitor:811 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 04:03:09,979 DEBUG SystemMonitor:811 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 04:03:09,979 INFO HandlerThread:811 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 04:03:09,981 INFO HandlerThread:811 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 04:03:09,981 INFO HandlerThread:811 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 04:03:09,981 INFO HandlerThread:811 [interfaces.py:finish():200] Joined network monitor +2024-05-23 04:03:09,981 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:09,982 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 04:03:09,982 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 3 +2024-05-23 04:03:09,982 DEBUG SenderThread:811 [sender.py:send():378] send: stats +2024-05-23 04:03:09,982 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:09,983 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 04:03:09,983 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:09,983 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 04:03:09,983 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 4 +2024-05-23 04:03:09,983 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:09,983 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 04:03:09,983 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:09,983 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 04:03:09,983 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 5 +2024-05-23 04:03:09,984 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:09,984 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 04:03:09,984 DEBUG SenderThread:811 [sender.py:send():378] send: summary +2024-05-23 04:03:09,985 INFO SenderThread:811 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 04:03:09,985 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:09,985 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 04:03:09,985 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 6 +2024-05-23 04:03:09,985 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:09,985 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 04:03:09,985 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:09,985 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 04:03:09,990 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 04:03:10,085 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 7 +2024-05-23 04:03:10,085 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:10,085 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 04:03:10,085 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:10,085 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 04:03:10,807 INFO Thread-12 :811 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/config.yaml +2024-05-23 04:03:10,807 INFO Thread-12 :811 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-summary.json +2024-05-23 04:03:10,975 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 04:03:12,386 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 8 +2024-05-23 04:03:12,386 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 04:03:12,386 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:12,386 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 04:03:12,386 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:12,386 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 04:03:12,386 INFO SenderThread:811 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 04:03:12,387 INFO SenderThread:811 [job_builder.py:_get_source_type():576] no source found +2024-05-23 04:03:12,387 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 9 +2024-05-23 04:03:12,387 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:12,387 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 04:03:12,387 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:12,387 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 04:03:12,387 INFO SenderThread:811 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 04:03:12,809 INFO SenderThread:811 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log +2024-05-23 04:03:12,809 INFO SenderThread:811 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files +2024-05-23 04:03:12,809 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-metadata.json wandb-metadata.json +2024-05-23 04:03:12,809 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-summary.json wandb-summary.json +2024-05-23 04:03:12,809 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log output.log +2024-05-23 04:03:12,812 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/requirements.txt requirements.txt +2024-05-23 04:03:12,812 INFO SenderThread:811 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/config.yaml config.yaml +2024-05-23 04:03:12,812 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 10 +2024-05-23 04:03:12,812 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:12,813 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 04:03:12,813 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:12,813 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 04:03:12,813 INFO SenderThread:811 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 04:03:12,976 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 04:03:12,976 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 04:03:13,078 INFO wandb-upload_0:811 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/wandb-summary.json +2024-05-23 04:03:13,371 INFO wandb-upload_1:811 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/output.log +2024-05-23 04:03:13,378 INFO wandb-upload_2:811 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/requirements.txt +2024-05-23 04:03:13,414 INFO wandb-upload_3:811 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/files/config.yaml +2024-05-23 04:03:13,614 INFO Thread-11 (_thread_body):811 [sender.py:transition_state():613] send defer: 11 +2024-05-23 04:03:13,614 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:13,614 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 04:03:13,615 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:13,615 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 04:03:13,615 INFO SenderThread:811 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 04:03:13,615 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 12 +2024-05-23 04:03:13,615 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:13,615 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 04:03:13,615 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:13,615 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 04:03:13,615 INFO SenderThread:811 [file_stream.py:finish():601] file stream finish called +2024-05-23 04:03:13,849 INFO SenderThread:811 [file_stream.py:finish():605] file stream finish is done +2024-05-23 04:03:13,849 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 13 +2024-05-23 04:03:13,849 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:13,849 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 04:03:13,849 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:13,849 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 04:03:13,849 INFO SenderThread:811 [sender.py:transition_state():613] send defer: 14 +2024-05-23 04:03:13,849 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: defer +2024-05-23 04:03:13,849 INFO HandlerThread:811 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 04:03:13,849 DEBUG SenderThread:811 [sender.py:send():378] send: final +2024-05-23 04:03:13,849 DEBUG SenderThread:811 [sender.py:send():378] send: footer +2024-05-23 04:03:13,849 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: defer +2024-05-23 04:03:13,850 INFO SenderThread:811 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 04:03:13,850 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 04:03:13,850 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 04:03:13,850 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 04:03:13,850 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 04:03:13,851 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 04:03:13,851 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 04:03:13,851 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 04:03:13,851 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 04:03:13,851 DEBUG SenderThread:811 [sender.py:send_request():405] send_request: server_info +2024-05-23 04:03:13,916 INFO MainThread:811 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 04:03:13,916 INFO MainThread:811 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 04:03:13,916 INFO MainThread:811 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 04:03:13,917 DEBUG HandlerThread:811 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 04:03:13,917 INFO HandlerThread:811 [handler.py:finish():882] shutting down handler +2024-05-23 04:03:14,851 INFO WriterThread:811 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/run-0kj3nw8j.wandb +2024-05-23 04:03:14,916 INFO SenderThread:811 [sender.py:finish():1545] shutting down sender +2024-05-23 04:03:14,916 INFO SenderThread:811 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 04:03:14,916 INFO SenderThread:811 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..4c41ce98339441d5570d200b5a1fa758ca93a70e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_setup.py:_flush():76] Configure stats pid to 655 +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 04:02:57,464 WARNING MainThread:655 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/logs/debug.log +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/logs/debug-internal.log +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_init.py:init():560] calling init triggers +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_init.py:init():610] starting backend +2024-05-23 04:02:57,464 INFO MainThread:655 [wandb_init.py:init():614] setting up manager +2024-05-23 04:02:57,468 INFO MainThread:655 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 04:02:57,469 INFO MainThread:655 [wandb_init.py:init():622] backend started and connected +2024-05-23 04:02:57,473 INFO MainThread:655 [wandb_init.py:init():711] updated telemetry +2024-05-23 04:02:57,483 INFO MainThread:655 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 04:02:57,795 INFO MainThread:655 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 04:02:57,909 INFO MainThread:655 [wandb_run.py:_on_init():2405] got version response +2024-05-23 04:02:57,909 INFO MainThread:655 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 04:02:58,198 INFO MainThread:655 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 04:02:58,198 INFO MainThread:655 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 04:02:58,199 INFO MainThread:655 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 04:02:58,199 INFO MainThread:655 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 04:02:58,202 INFO MainThread:655 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 04:03:14,918 WARNING MsgRouterThr:655 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/run-0kj3nw8j.wandb b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/run-0kj3nw8j.wandb new file mode 100644 index 0000000000000000000000000000000000000000..4dcc28325241c5d23f3a4701638765391fc6dfb0 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_040257-0kj3nw8j/run-0kj3nw8j.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3dc4eac94b70a8b012fc7aee3e7f12d6c3589a1b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716444322 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..f859f3242b65abdcb01c49596032379694f74639 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:06:05:23,265 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:06:05:32,634 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:06:05:32,636 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:06:05:32,636 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100'} +2024-05-23:06:05:34,955 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..941a8e5de19c9ca9b312950ad3d50aa45ac3d6fd --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T06:05:23.042386", + "startedAt": "2024-05-23T06:05:22.550846", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2334.2706937499997, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 209.5741081237793 + } + }, + "memory": { + "total": 1007.4379425048828 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..018b4068e81a342faa2a3b0691dee6965106bcc3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 12}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..114f511525756482ca8143b628d84d0a8bf007ad --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/logs/debug-internal.log @@ -0,0 +1,185 @@ +2024-05-23 06:05:22,574 INFO StreamThr :825 [internal.py:wandb_internal():85] W&B internal server running at pid: 825, started at: 2024-05-23 06:05:22.572275 +2024-05-23 06:05:22,579 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status +2024-05-23 06:05:22,580 INFO WriterThread:825 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/run-0i9x5bzp.wandb +2024-05-23 06:05:22,582 DEBUG SenderThread:825 [sender.py:send():378] send: header +2024-05-23 06:05:22,586 DEBUG SenderThread:825 [sender.py:send():378] send: run +2024-05-23 06:05:22,845 INFO SenderThread:825 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files +2024-05-23 06:05:22,846 INFO SenderThread:825 [sender.py:_start_run_threads():1123] run started: 0i9x5bzp with start time 1716444322.573536 +2024-05-23 06:05:22,849 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 06:05:22,849 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: check_version +2024-05-23 06:05:22,965 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 06:05:22,968 DEBUG HandlerThread:825 [system_info.py:__init__():26] System info init +2024-05-23 06:05:22,968 DEBUG HandlerThread:825 [system_info.py:__init__():41] System info init done +2024-05-23 06:05:22,968 INFO HandlerThread:825 [system_monitor.py:start():194] Starting system monitor +2024-05-23 06:05:22,968 INFO SystemMonitor:825 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 06:05:22,968 INFO HandlerThread:825 [system_monitor.py:probe():214] Collecting system info +2024-05-23 06:05:22,975 INFO SystemMonitor:825 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 06:05:22,976 INFO SystemMonitor:825 [interfaces.py:start():188] Started disk monitoring +2024-05-23 06:05:22,976 INFO SystemMonitor:825 [interfaces.py:start():188] Started memory monitoring +2024-05-23 06:05:22,978 INFO SystemMonitor:825 [interfaces.py:start():188] Started network monitoring +2024-05-23 06:05:23,042 DEBUG HandlerThread:825 [system_info.py:probe():150] Probing system +2024-05-23 06:05:23,045 DEBUG HandlerThread:825 [system_info.py:_probe_git():135] Probing git +2024-05-23 06:05:23,055 ERROR HandlerThread:825 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 06:05:23,055 DEBUG HandlerThread:825 [system_info.py:_probe_git():143] Probing git done +2024-05-23 06:05:23,056 DEBUG HandlerThread:825 [system_info.py:probe():198] Probing system done +2024-05-23 06:05:23,056 DEBUG HandlerThread:825 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T06:05:23.042386', 'startedAt': '2024-05-23T06:05:22.550846', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2334.2706937499997, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 209.5741081237793}}, 'memory': {'total': 1007.4379425048828}} +2024-05-23 06:05:23,056 INFO HandlerThread:825 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 06:05:23,056 INFO HandlerThread:825 [system_monitor.py:probe():227] Publishing system info +2024-05-23 06:05:23,059 INFO HandlerThread:825 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 06:05:23,066 DEBUG SenderThread:825 [sender.py:send():378] send: files +2024-05-23 06:05:23,066 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 06:05:23,248 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 06:05:23,248 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: python_packages +2024-05-23 06:05:23,249 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 06:05:23,251 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: stop_status +2024-05-23 06:05:23,381 DEBUG SenderThread:825 [sender.py:send():378] send: telemetry +2024-05-23 06:05:23,711 INFO wandb-upload_0:825 [upload_job.py:push():130] Uploaded file /tmp/tmp_8k8hei3wandb/xkgmu84p-wandb-metadata.json +2024-05-23 06:05:23,848 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-metadata.json +2024-05-23 06:05:23,848 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log +2024-05-23 06:05:23,848 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/requirements.txt +2024-05-23 06:05:25,848 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log +2024-05-23 06:05:28,383 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:05:33,637 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:05:33,856 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log +2024-05-23 06:05:34,973 DEBUG SenderThread:825 [sender.py:send():378] send: exit +2024-05-23 06:05:34,973 INFO SenderThread:825 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 06:05:34,973 INFO SenderThread:825 [sender.py:send_exit():587] handling runtime: 12 +2024-05-23 06:05:34,978 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 06:05:34,978 INFO SenderThread:825 [sender.py:send_exit():593] send defer +2024-05-23 06:05:34,978 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:34,978 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 06:05:34,978 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:34,978 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 06:05:34,978 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 1 +2024-05-23 06:05:34,978 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:34,979 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 06:05:34,979 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:34,979 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 06:05:34,979 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 2 +2024-05-23 06:05:34,979 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:34,979 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 06:05:34,979 INFO HandlerThread:825 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 06:05:34,979 DEBUG SystemMonitor:825 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 06:05:34,979 DEBUG SystemMonitor:825 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 06:05:34,980 DEBUG SystemMonitor:825 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 06:05:34,980 INFO HandlerThread:825 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 06:05:34,982 INFO HandlerThread:825 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 06:05:34,982 INFO HandlerThread:825 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 06:05:34,982 INFO HandlerThread:825 [interfaces.py:finish():200] Joined network monitor +2024-05-23 06:05:34,983 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:34,983 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 06:05:34,983 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 3 +2024-05-23 06:05:34,983 DEBUG SenderThread:825 [sender.py:send():378] send: stats +2024-05-23 06:05:34,984 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:34,984 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 06:05:34,984 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:34,984 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 06:05:34,984 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 4 +2024-05-23 06:05:34,985 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:34,985 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 06:05:34,985 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:34,985 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 06:05:34,985 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 5 +2024-05-23 06:05:34,985 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:34,985 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 06:05:34,985 DEBUG SenderThread:825 [sender.py:send():378] send: summary +2024-05-23 06:05:34,986 INFO SenderThread:825 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 06:05:34,986 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:34,986 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 06:05:34,986 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 6 +2024-05-23 06:05:34,986 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:34,986 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 06:05:34,986 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:34,986 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 06:05:34,991 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 06:05:35,067 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 7 +2024-05-23 06:05:35,067 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:35,067 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 06:05:35,067 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:35,067 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 06:05:35,858 INFO Thread-12 :825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/config.yaml +2024-05-23 06:05:35,858 INFO Thread-12 :825 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-summary.json +2024-05-23 06:05:35,973 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:05:37,409 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 8 +2024-05-23 06:05:37,410 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:05:37,410 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:37,410 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 06:05:37,410 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:37,410 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 06:05:37,410 INFO SenderThread:825 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 06:05:37,411 INFO SenderThread:825 [job_builder.py:_get_source_type():576] no source found +2024-05-23 06:05:37,411 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 9 +2024-05-23 06:05:37,411 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:37,411 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 06:05:37,411 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:37,411 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 06:05:37,411 INFO SenderThread:825 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 06:05:37,860 INFO SenderThread:825 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log +2024-05-23 06:05:37,861 INFO SenderThread:825 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files +2024-05-23 06:05:37,861 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-summary.json wandb-summary.json +2024-05-23 06:05:37,861 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log output.log +2024-05-23 06:05:37,863 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/config.yaml config.yaml +2024-05-23 06:05:37,864 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-metadata.json wandb-metadata.json +2024-05-23 06:05:37,864 INFO SenderThread:825 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/requirements.txt requirements.txt +2024-05-23 06:05:37,864 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 10 +2024-05-23 06:05:37,864 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:37,864 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 06:05:37,864 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:37,864 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 06:05:37,864 INFO SenderThread:825 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 06:05:37,973 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:05:37,973 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:05:38,114 INFO wandb-upload_0:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/wandb-summary.json +2024-05-23 06:05:38,521 INFO wandb-upload_3:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/requirements.txt +2024-05-23 06:05:38,530 INFO wandb-upload_1:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/output.log +2024-05-23 06:05:38,564 INFO wandb-upload_2:825 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/files/config.yaml +2024-05-23 06:05:38,765 INFO Thread-11 (_thread_body):825 [sender.py:transition_state():613] send defer: 11 +2024-05-23 06:05:38,765 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:38,765 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 06:05:38,765 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:38,765 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 06:05:38,765 INFO SenderThread:825 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 06:05:38,765 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 12 +2024-05-23 06:05:38,765 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:38,766 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 06:05:38,766 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:38,766 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 06:05:38,766 INFO SenderThread:825 [file_stream.py:finish():601] file stream finish called +2024-05-23 06:05:38,974 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:05:38,980 INFO SenderThread:825 [file_stream.py:finish():605] file stream finish is done +2024-05-23 06:05:38,981 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 13 +2024-05-23 06:05:38,981 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:05:38,981 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:38,981 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 06:05:38,981 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:38,981 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 06:05:38,981 INFO SenderThread:825 [sender.py:transition_state():613] send defer: 14 +2024-05-23 06:05:38,981 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: defer +2024-05-23 06:05:38,981 INFO HandlerThread:825 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 06:05:38,981 DEBUG SenderThread:825 [sender.py:send():378] send: final +2024-05-23 06:05:38,981 DEBUG SenderThread:825 [sender.py:send():378] send: footer +2024-05-23 06:05:38,982 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: defer +2024-05-23 06:05:38,982 INFO SenderThread:825 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 06:05:38,982 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:05:38,982 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 06:05:38,982 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 06:05:38,983 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 06:05:38,983 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 06:05:38,983 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 06:05:38,983 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:05:38,983 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 06:05:38,983 DEBUG SenderThread:825 [sender.py:send_request():405] send_request: server_info +2024-05-23 06:05:39,044 INFO MainThread:825 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 06:05:39,044 INFO MainThread:825 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 06:05:39,044 INFO MainThread:825 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 06:05:39,045 DEBUG HandlerThread:825 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 06:05:39,045 INFO HandlerThread:825 [handler.py:finish():882] shutting down handler +2024-05-23 06:05:39,983 INFO WriterThread:825 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/run-0i9x5bzp.wandb +2024-05-23 06:05:40,044 INFO SenderThread:825 [sender.py:finish():1545] shutting down sender +2024-05-23 06:05:40,044 INFO SenderThread:825 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 06:05:40,044 INFO SenderThread:825 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..66649d85462871794c20aaea5231d2ab1480cd2d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_setup.py:_flush():76] Configure stats pid to 669 +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 06:05:22,567 WARNING MainThread:669 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/logs/debug.log +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/logs/debug-internal.log +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_init.py:init():560] calling init triggers +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_init.py:init():610] starting backend +2024-05-23 06:05:22,567 INFO MainThread:669 [wandb_init.py:init():614] setting up manager +2024-05-23 06:05:22,571 INFO MainThread:669 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 06:05:22,573 INFO MainThread:669 [wandb_init.py:init():622] backend started and connected +2024-05-23 06:05:22,577 INFO MainThread:669 [wandb_init.py:init():711] updated telemetry +2024-05-23 06:05:22,585 INFO MainThread:669 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 06:05:22,848 INFO MainThread:669 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 06:05:22,959 INFO MainThread:669 [wandb_run.py:_on_init():2405] got version response +2024-05-23 06:05:22,959 INFO MainThread:669 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 06:05:23,249 INFO MainThread:669 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 06:05:23,249 INFO MainThread:669 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 06:05:23,249 INFO MainThread:669 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 06:05:23,249 INFO MainThread:669 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 06:05:23,262 INFO MainThread:669 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 06:05:40,045 WARNING MsgRouterThr:669 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/run-0i9x5bzp.wandb b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/run-0i9x5bzp.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ffc10e2eddc27ab8d99fe744bfee41f3b9637ee4 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_060522-0i9x5bzp/run-0i9x5bzp.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..08f3fdee2e79b9b07c1b8bea2a2d405d30b15f64 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/logs/debug-internal.log @@ -0,0 +1,1590 @@ +2024-05-29 13:05:47,767 INFO StreamThr :899 [internal.py:wandb_internal():85] W&B internal server running at pid: 899, started at: 2024-05-29 13:05:47.765019 +2024-05-29 13:05:47,771 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status +2024-05-29 13:05:47,772 INFO WriterThread:899 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/run-x16dn3vu.wandb +2024-05-29 13:05:47,774 DEBUG SenderThread:899 [sender.py:send():378] send: header +2024-05-29 13:05:47,781 DEBUG SenderThread:899 [sender.py:send():378] send: run +2024-05-29 13:05:48,044 INFO SenderThread:899 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files +2024-05-29 13:05:48,044 INFO SenderThread:899 [sender.py:_start_run_threads():1123] run started: x16dn3vu with start time 1716987947.765084 +2024-05-29 13:05:48,048 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: check_version +2024-05-29 13:05:48,048 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: check_version +2024-05-29 13:05:48,162 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: run_start +2024-05-29 13:05:48,164 DEBUG HandlerThread:899 [system_info.py:__init__():26] System info init +2024-05-29 13:05:48,164 DEBUG HandlerThread:899 [system_info.py:__init__():41] System info init done +2024-05-29 13:05:48,164 INFO HandlerThread:899 [system_monitor.py:start():194] Starting system monitor +2024-05-29 13:05:48,164 INFO SystemMonitor:899 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-29 13:05:48,164 INFO HandlerThread:899 [system_monitor.py:probe():214] Collecting system info +2024-05-29 13:05:48,171 INFO SystemMonitor:899 [interfaces.py:start():188] Started cpu monitoring +2024-05-29 13:05:48,172 INFO SystemMonitor:899 [interfaces.py:start():188] Started disk monitoring +2024-05-29 13:05:48,172 INFO SystemMonitor:899 [interfaces.py:start():188] Started memory monitoring +2024-05-29 13:05:48,177 INFO SystemMonitor:899 [interfaces.py:start():188] Started network monitoring +2024-05-29 13:05:48,238 DEBUG HandlerThread:899 [system_info.py:probe():150] Probing system +2024-05-29 13:05:48,242 DEBUG HandlerThread:899 [system_info.py:_probe_git():135] Probing git +2024-05-29 13:05:48,252 ERROR HandlerThread:899 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-29 13:05:48,252 DEBUG HandlerThread:899 [system_info.py:_probe_git():143] Probing git done +2024-05-29 13:05:48,252 DEBUG HandlerThread:899 [system_info.py:probe():198] Probing system done +2024-05-29 13:05:48,252 DEBUG HandlerThread:899 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-29T13:05:48.238638', 'startedAt': '2024-05-29T13:05:47.742131', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint/llamav2-3b/hf/global_step40000,tokenizer=huggyllama/llama-13b', '--tasks', 'winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq', '--batch_size', 'auto', '--wandb_args', 'project=english-eval,group=exp1'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-20-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.3872875, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3386.857, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.81209182739258}}, 'memory': {'total': 1007.437931060791}} +2024-05-29 13:05:48,252 INFO HandlerThread:899 [system_monitor.py:probe():224] Finished collecting system info +2024-05-29 13:05:48,252 INFO HandlerThread:899 [system_monitor.py:probe():227] Publishing system info +2024-05-29 13:05:48,255 INFO HandlerThread:899 [system_monitor.py:probe():229] Finished publishing system info +2024-05-29 13:05:48,261 DEBUG SenderThread:899 [sender.py:send():378] send: files +2024-05-29 13:05:48,261 INFO SenderThread:899 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-29 13:05:48,436 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: python_packages +2024-05-29 13:05:48,436 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: python_packages +2024-05-29 13:05:48,439 DEBUG SenderThread:899 [sender.py:send():378] send: telemetry +2024-05-29 13:05:48,439 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:05:48,439 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:05:48,809 INFO wandb-upload_0:899 [upload_job.py:push():130] Uploaded file /tmp/tmp0s4o2ndwwandb/79r29nnz-wandb-metadata.json +2024-05-29 13:05:49,047 INFO Thread-12 :899 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/requirements.txt +2024-05-29 13:05:49,047 INFO Thread-12 :899 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/wandb-metadata.json +2024-05-29 13:05:49,047 INFO Thread-12 :899 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:05:51,046 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:05:53,521 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:05:59,022 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:05:59,053 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:03,078 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:03,438 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:06:03,438 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:06:04,562 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:09,563 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:14,563 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:18,438 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:06:18,438 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:06:20,516 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:21,147 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/config.yaml +2024-05-29 13:06:25,605 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:26,353 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:28,608 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:30,611 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:32,292 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:32,628 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:33,442 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:06:33,443 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:06:34,767 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:36,770 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:37,305 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:38,772 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:40,774 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:42,443 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:42,777 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:44,825 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:46,967 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:47,454 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:48,177 DEBUG SystemMonitor:899 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-29 13:06:48,180 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:06:48,443 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:06:48,444 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:06:49,146 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:51,148 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:53,211 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:53,428 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:55,436 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:57,474 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:06:58,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:06:59,608 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:07:00,774 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:07:03,021 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:07:03,448 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:07:03,448 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:07:03,551 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:05,311 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:07:07,526 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:07:08,814 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:09,034 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:07:13,814 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:18,182 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:07:18,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:07:18,471 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:07:18,917 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:18,977 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:07:21,159 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:07:23,918 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:28,918 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:33,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:07:33,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:07:34,544 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:39,544 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:44,545 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:48,185 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:07:48,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:07:48,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:07:50,531 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:07:55,532 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:00,533 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:03,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:08:03,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:08:05,645 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:10,645 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:15,646 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:18,188 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:08:18,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:08:18,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:08:21,567 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:26,568 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:31,569 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:33,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:08:33,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:08:36,652 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:41,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:46,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:48,189 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:08:48,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:08:48,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:08:51,114 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:08:52,347 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:08:53,471 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:08:57,347 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:02,544 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:03,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:09:03,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:09:03,670 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:09:07,643 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:12,644 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:16,838 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:09:17,837 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:18,191 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:09:18,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:09:18,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:09:23,583 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:28,584 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:33,062 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:09:33,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:09:33,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:09:33,651 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:38,652 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:41,639 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:09:43,972 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:48,194 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:09:48,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:09:48,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:09:49,592 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:53,635 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:09:54,643 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:09:59,644 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:03,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:10:03,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:10:05,002 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:10:05,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:07,667 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:10:10,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:15,819 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:16,835 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:10:18,196 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:10:18,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:10:18,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:10:21,560 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:26,636 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:26,849 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:10:29,099 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:10:32,164 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:33,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:10:33,472 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:10:37,570 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:42,708 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:43,122 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:10:47,715 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:48,198 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:10:48,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:10:48,473 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:10:51,766 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:10:53,150 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:10:58,151 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:03,427 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:11:03,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:11:03,473 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:11:03,558 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:08,593 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:10,194 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:11:11,195 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:11:13,898 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:18,200 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:11:18,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:11:18,473 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:11:19,630 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:23,259 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:11:25,063 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:27,028 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:11:29,661 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:11:31,029 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:33,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:11:33,473 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:11:36,563 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:41,564 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:43,678 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:11:46,643 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:47,038 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:11:48,202 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:11:48,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:11:48,473 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:11:52,600 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:11:57,601 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:03,364 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:03,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:12:03,473 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:12:04,025 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:12:05,443 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:12:09,337 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:14,338 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:17,763 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:12:18,204 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:12:18,472 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:12:18,473 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:12:19,556 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:23,122 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:12:25,494 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:30,495 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:33,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:12:33,473 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:12:36,211 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:37,539 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:12:41,395 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:41,630 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:12:46,396 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:48,207 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:12:48,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:12:48,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:12:51,619 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:55,768 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:12:56,898 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:12:59,078 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:13:02,792 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:03,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:13:03,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:13:08,543 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:13,707 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:15,971 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:13:17,011 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:13:18,209 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:13:18,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:13:18,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:13:19,633 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:24,634 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:29,634 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:33,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:13:33,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:13:33,577 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:13:35,577 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:40,578 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:45,951 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:46,111 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:13:48,211 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:13:48,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:13:48,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:13:49,886 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:13:51,540 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:13:52,202 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:13:56,541 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:01,541 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:03,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:14:03,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:14:06,038 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:14:06,899 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:07,130 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:14:11,900 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:16,901 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:18,213 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:14:18,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:14:18,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:14:22,637 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:23,330 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:14:27,637 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:32,638 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:33,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:14:33,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:14:36,162 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:14:37,655 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:39,578 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:14:42,789 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:47,790 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:48,216 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:14:48,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:14:48,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:14:52,809 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:14:55,462 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:14:58,756 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:03,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:15:03,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:15:04,656 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:08,386 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:15:09,678 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:12,005 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:15:14,894 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:18,217 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:15:18,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:15:18,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:15:20,571 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:25,898 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:15:26,123 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:31,124 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:33,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:15:33,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:15:36,420 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:38,291 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:15:39,346 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:15:41,604 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:46,605 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:48,219 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:15:48,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:15:48,474 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:15:52,210 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:15:53,443 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:15:56,194 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:15:58,129 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:03,129 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:03,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:16:03,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:16:06,195 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:16:08,205 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:09,827 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:16:13,522 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:18,222 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:16:18,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:16:18,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:16:18,587 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:23,859 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:16:24,516 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:26,294 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:16:29,517 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:33,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:16:33,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:16:34,639 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:37,679 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:16:39,959 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:40,061 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:16:44,960 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:48,224 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:16:48,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:16:48,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:16:50,607 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:53,778 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:16:56,042 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:16:56,182 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:01,043 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:03,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:17:03,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:17:06,502 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:06,576 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:10,225 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:12,269 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:17,270 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:18,226 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:17:18,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:17:18,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:17:22,342 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:23,939 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:27,756 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:32,757 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:33,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:17:33,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:17:36,116 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:38,287 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:38,414 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:39,415 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:43,287 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:48,228 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:17:48,474 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:17:48,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:17:48,643 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:51,894 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:53,644 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:17:54,286 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:17:58,644 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:03,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:18:03,475 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:18:03,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:05,975 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:18:08,097 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:18:09,084 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:14,084 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:18,230 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:18:18,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:18:18,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:18:19,338 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:22,079 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:18:25,133 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:30,134 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:33,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:18:33,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:18:35,612 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:36,258 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:18:40,613 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:45,613 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:47,723 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:18:48,232 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:18:48,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:18:48,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:18:50,262 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:18:50,632 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:18:55,632 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:00,655 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:19:01,558 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:01,706 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:19:03,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:19:03,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:19:06,634 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:11,710 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:13,864 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:19:17,126 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:18,235 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:19:18,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:19:18,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:19:22,589 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:26,347 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:19:27,623 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:32,624 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:33,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:19:33,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:19:38,139 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:38,535 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:19:39,736 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:19:43,139 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:48,140 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:48,237 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:19:48,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:19:48,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:19:50,379 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:19:52,763 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:19:53,709 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:19:58,711 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:03,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:20:03,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:20:03,562 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:20:04,647 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:09,647 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:14,746 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:16,234 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:20:18,239 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:20:18,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:20:18,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:20:20,613 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:25,613 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:28,802 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:20:31,181 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:33,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:20:33,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:20:36,633 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:39,831 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:20:42,314 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:20:42,391 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:47,392 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:48,241 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:20:48,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:20:48,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:20:52,400 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:20:52,494 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:20:53,554 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:20:57,401 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:02,402 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:03,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:21:03,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:21:04,486 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:21:07,066 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:21:07,607 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:12,608 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:16,131 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:21:18,243 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:21:18,246 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:18,350 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:21:18,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:21:18,476 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:21:23,601 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:27,690 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:21:29,243 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:33,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:21:33,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:21:34,576 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:40,254 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:21:40,522 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:45,522 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:48,245 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:21:48,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:21:48,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:21:49,618 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:21:50,837 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:21:52,570 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:21:55,837 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:01,823 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:02,541 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:22:03,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:22:03,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:22:07,642 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:12,162 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:22:12,829 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:14,782 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:22:17,829 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:18,248 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:22:18,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:22:18,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:22:22,846 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:24,387 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:22:27,846 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:32,847 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:33,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:22:33,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:22:36,111 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:22:37,966 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:42,967 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:46,241 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:22:48,249 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:22:48,252 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:48,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:22:48,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:22:53,586 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:22:57,146 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:22:58,302 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:22:58,605 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:03,476 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:23:03,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:23:03,637 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:07,731 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:23:09,072 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:14,073 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:17,955 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:23:18,251 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:23:18,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:23:18,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:23:19,624 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:24,625 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:27,843 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:23:30,041 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:33,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:23:33,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:23:35,579 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:38,602 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:23:40,798 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:23:40,818 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:45,818 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:48,254 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:23:48,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:23:48,477 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:23:49,890 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:23:51,579 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:56,579 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:23:59,810 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:24:02,176 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:03,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:24:03,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:24:07,559 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:10,330 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:24:12,770 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:24:12,838 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:17,838 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:18,256 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:24:18,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:24:18,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:24:22,583 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:24:23,500 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:28,500 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:32,326 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:24:33,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:24:33,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:24:33,583 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:38,584 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:42,698 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:24:43,997 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:48,258 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:24:48,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:24:48,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:24:49,566 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:24:51,979 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:24:55,071 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:00,072 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:02,339 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:03,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:25:03,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:25:05,649 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:10,649 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:13,094 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:16,319 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:18,260 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:25:18,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:25:18,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:25:21,412 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:22,059 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:26,413 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:31,418 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:32,642 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:33,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:25:33,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:25:34,667 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:36,624 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:41,697 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:42,336 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:44,638 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:46,698 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:48,262 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:25:48,477 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:25:48,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:25:52,323 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:52,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:25:54,567 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:25:57,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:01,954 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:26:03,399 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:03,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:26:03,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:26:08,598 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:10,116 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:26:14,564 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:18,265 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:26:18,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:26:18,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:26:19,645 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:21,126 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:26:24,646 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:28,419 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:26:30,389 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:33,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:26:33,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:26:35,609 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:38,326 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:26:40,646 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:45,657 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:46,183 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:26:48,267 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:26:48,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:26:48,478 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:26:51,612 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:26:56,023 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:26:57,556 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:02,557 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:03,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:27:03,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:27:04,172 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:27:07,594 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:13,394 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:14,067 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:27:18,269 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:27:18,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:27:18,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:27:18,615 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:22,722 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:27:24,198 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:29,199 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:30,695 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:27:33,114 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:27:33,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:27:33,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:27:34,594 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:40,370 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:40,546 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:27:45,370 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:48,271 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:27:48,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:27:48,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:27:50,167 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:27:50,602 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:55,603 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:27:58,278 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:01,092 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:03,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:28:03,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:28:07,055 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:07,182 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:08,183 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:12,056 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:17,586 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:17,744 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:18,273 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:28:18,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:28:18,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:28:23,134 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:25,332 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:28,659 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:33,478 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:28:33,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:28:34,561 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:35,210 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:39,562 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:42,818 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:45,329 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:48,275 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:28:48,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:28:48,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:28:50,663 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:50,946 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:28:52,855 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:28:55,947 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:00,257 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:29:01,873 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:03,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:29:03,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:29:06,902 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:08,324 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:29:12,213 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:16,230 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:29:17,841 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:18,277 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:29:18,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:29:18,479 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:29:23,640 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:25,260 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:29:26,274 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:29:29,248 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:33,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:29:33,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:29:34,512 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:29:34,611 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:39,612 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:42,200 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:29:45,412 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:48,280 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:29:48,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:29:48,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:29:50,275 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:29:50,708 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:29:55,709 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:01,027 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:01,540 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:03,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:30:03,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:30:06,462 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:07,126 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:08,762 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:12,127 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:15,043 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:17,166 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:17,186 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:18,282 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:30:18,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:30:18,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:30:22,502 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:22,688 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:27,694 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:30,898 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:33,127 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:33,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:30:33,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:30:38,318 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:38,330 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:43,330 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:46,745 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:48,284 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:30:48,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:30:48,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:30:48,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:53,634 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:54,171 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:30:55,006 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:30:59,172 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:00,300 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:03,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:31:03,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:31:04,813 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:08,266 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:10,509 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:16,083 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:16,783 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:18,286 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:31:18,479 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:31:18,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:31:21,640 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:25,531 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:26,804 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:30,271 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:32,374 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:33,443 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:33,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:31:33,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:31:37,397 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:38,447 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:42,813 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:47,096 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:48,288 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:31:48,291 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:48,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:31:48,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:31:53,865 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:31:54,596 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:31:58,866 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:01,171 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:03,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:32:03,480 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:32:03,719 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:04,579 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:08,365 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:09,596 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:14,596 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:17,042 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:18,291 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:32:18,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:32:18,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:32:19,598 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:24,715 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:24,815 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:29,716 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:30,595 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:33,198 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:33,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:32:33,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:32:35,625 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:38,378 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:40,817 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:46,238 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:46,967 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:48,293 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:32:48,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:32:48,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:32:51,549 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:32:54,687 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:32:56,702 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:01,380 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:33:02,041 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:03,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:33:03,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:33:07,161 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:09,367 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:33:12,766 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:16,780 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:33:18,187 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:18,295 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:33:18,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:33:18,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:33:23,587 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:24,716 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:33:28,587 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:31,118 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:33:33,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:33:33,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:33:33,610 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:38,533 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:33:38,952 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:43,953 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:46,950 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:33:48,297 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:33:48,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:33:48,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:33:49,616 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:53,370 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:33:55,446 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:33:55,594 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:00,429 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:00,583 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:03,480 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:34:03,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:34:05,599 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:08,966 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:11,336 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:15,364 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:16,559 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:17,827 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:18,299 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:34:18,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:34:18,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:34:21,591 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:22,906 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:26,800 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:31,121 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:31,964 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:33,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:34:33,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:34:37,238 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:37,339 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:39,543 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:42,339 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:44,638 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:47,629 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:48,301 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:34:48,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:34:48,481 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:34:53,230 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:34:53,612 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:58,650 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:34:59,347 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:03,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:35:03,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:35:03,651 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:05,596 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:06,603 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:08,695 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:12,582 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:14,255 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:18,303 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:35:18,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:35:18,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:35:19,311 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:20,984 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:24,727 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:29,570 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:29,844 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:33,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:35:33,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:35:34,862 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:34,905 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:39,906 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:42,875 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:45,079 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:48,306 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:35:48,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:35:48,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:35:49,214 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:50,715 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:51,562 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:35:55,715 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:35:57,154 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:01,215 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:03,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:36:03,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:36:05,344 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:06,566 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:11,586 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:12,410 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:12,587 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:17,491 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:18,308 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:36:18,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:36:18,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:36:19,217 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:22,555 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:25,692 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:26,727 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:27,803 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:32,775 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:33,145 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:33,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:36:33,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:36:38,572 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:41,156 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:44,305 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:47,531 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:48,310 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:36:48,481 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:36:48,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:36:49,629 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:36:54,866 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:36:55,432 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:00,488 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:01,626 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:02,642 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:03,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:37:03,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:37:05,647 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:09,647 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:11,056 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:14,770 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:16,307 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:18,312 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:37:18,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:37:18,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:37:20,843 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:21,562 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:23,198 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:26,563 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:29,779 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:31,883 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:33,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:37:33,482 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:37:34,782 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:37,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:41,117 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:43,499 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:46,843 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:48,314 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:37:48,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:37:48,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:37:48,634 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:53,210 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:53,645 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:37:55,664 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:37:58,646 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:01,353 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:03,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:38:03,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:38:04,545 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:06,826 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:10,081 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:12,907 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:15,678 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:15,914 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:18,316 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:38:18,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:38:18,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:38:20,986 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:21,327 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:26,612 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:27,391 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:31,746 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:33,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:38:33,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:38:33,558 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:37,611 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:38,758 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:42,988 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:44,746 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:48,055 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:48,318 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:38:48,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:38:48,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:38:50,850 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:53,544 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:57,187 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:38:58,908 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:38:59,492 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:03,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:39:03,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:39:04,821 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:05,728 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:09,885 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:11,750 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:14,993 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:17,663 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:18,321 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:39:18,482 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:39:18,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:39:20,630 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:23,006 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:26,424 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:29,763 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:31,624 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:33,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:39:33,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:39:35,846 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:36,847 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:36,963 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:42,307 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:43,424 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:47,355 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:48,323 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:39:48,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:39:48,483 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:39:49,313 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:52,597 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:39:55,444 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:39:58,595 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:01,531 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:03,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:40:03,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:40:04,569 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:06,785 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:08,829 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:10,114 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:15,230 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:15,329 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:18,325 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:40:18,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:40:18,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:40:20,496 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:21,128 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:25,502 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:27,546 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:30,958 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:33,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:40:33,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:40:33,876 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:36,652 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:38,998 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:42,505 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:45,731 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:47,662 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:48,327 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:40:48,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:40:48,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:40:50,900 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:52,791 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:53,436 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:40:57,833 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:40:59,058 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:03,056 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:03,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:41:03,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:41:05,106 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:08,202 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:11,214 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:13,705 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:17,371 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:18,330 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:41:18,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:41:18,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:41:19,634 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:23,866 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:24,919 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:25,004 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:30,284 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:31,093 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:33,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:41:33,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:41:35,311 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:37,046 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:40,413 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:43,074 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:45,432 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:48,331 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:41:48,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:41:48,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:41:49,436 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:50,749 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:55,655 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:41:56,004 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:41:57,826 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:01,038 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:01,665 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:03,483 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:42:03,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:42:03,919 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:06,627 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:08,194 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:09,443 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:11,819 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:15,689 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:16,860 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:18,333 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:42:18,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:42:18,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:42:22,146 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:22,400 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:23,147 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:27,438 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:29,446 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:32,563 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:33,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:42:33,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:42:35,860 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:37,756 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:41,162 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:43,192 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:47,550 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:48,336 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:42:48,339 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:48,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:42:48,484 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:42:49,875 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:53,539 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:42:55,162 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:42:58,742 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:01,788 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:03,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:43:03,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:43:04,620 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:08,064 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:10,366 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:13,058 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:15,417 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:18,338 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:43:18,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:43:18,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:43:19,718 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:20,840 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:25,755 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:26,200 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:28,254 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:31,294 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:33,050 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:33,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:43:33,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:43:36,566 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:39,485 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:41,881 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:45,851 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:47,125 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:48,340 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:43:48,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:43:48,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:43:51,099 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:52,551 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:58,199 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:43:58,358 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:43:59,235 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:03,450 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:03,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:44:03,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:44:05,399 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:08,580 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:11,828 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:13,746 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:17,869 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:18,342 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:44:18,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:44:18,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:44:19,571 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:24,270 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:24,676 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:29,554 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:30,303 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:32,326 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:33,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:44:33,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:44:35,351 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:36,259 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:37,774 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:40,471 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:44,246 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:45,479 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:45,766 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:48,344 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:44:48,484 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:44:48,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:44:49,259 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:51,419 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:44:51,520 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:56,668 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:44:57,311 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:02,517 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:03,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:45:03,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:45:03,633 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:08,406 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:10,038 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:14,140 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:15,976 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:18,346 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:45:18,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:45:18,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:45:19,862 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:21,837 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:25,215 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:25,618 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:30,705 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:31,286 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:33,344 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:33,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:45:33,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:45:36,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:39,972 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:41,949 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:43,528 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:47,612 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:48,348 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:45:48,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:45:48,485 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:45:50,168 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:53,598 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:45:56,090 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:45:59,386 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:01,150 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:03,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:46:03,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:46:04,571 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:07,699 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:09,949 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:13,380 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:15,750 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:18,350 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:46:18,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:46:18,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:46:19,364 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:21,606 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:26,043 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:27,473 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:32,200 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:33,230 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:33,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:46:33,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:46:38,204 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:38,362 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:43,203 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:43,787 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:48,352 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:46:48,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:46:48,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:46:48,860 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:49,424 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:51,948 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:54,081 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:46:55,369 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:57,780 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:46:59,215 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:03,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:47:03,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:47:04,107 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:04,633 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:07,618 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:09,748 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:10,619 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:16,372 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:16,501 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:18,355 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:47:18,485 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:47:18,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:47:21,684 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:22,115 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:27,276 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:27,971 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:33,249 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:33,486 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:47:33,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:47:34,215 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:37,962 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:38,791 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:40,398 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:44,567 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:46,438 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:48,357 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:47:48,486 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:47:48,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:47:50,403 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:51,257 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:47:56,287 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:47:57,491 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:01,379 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:01,635 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:03,486 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:48:03,487 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:48:04,326 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:06,389 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:08,182 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:11,475 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:13,444 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:16,885 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:18,359 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:48:18,486 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:48:18,486 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:48:20,598 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:22,642 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:25,435 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:28,302 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:30,647 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:31,878 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:33,486 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:48:33,487 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:48:33,606 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:36,554 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:37,894 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:39,067 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:41,598 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:44,516 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:47,663 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:48,361 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:48:48,486 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:48:48,487 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:48:49,611 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:54,539 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:48:55,406 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:48:57,954 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:49:00,840 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:03,486 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:49:03,487 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:49:03,933 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:49:06,588 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:10,278 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:49:11,682 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:16,682 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:18,364 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:49:18,614 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:49:18,614 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:49:21,724 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:26,724 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:31,725 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:33,545 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:49:33,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:49:35,799 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:49:36,831 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:41,832 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:46,833 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:48,367 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:49:48,546 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:49:48,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:49:52,673 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:49:57,674 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:02,675 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:03,545 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:50:03,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:50:08,632 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:13,633 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:18,369 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:50:18,545 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:50:18,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:50:18,689 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:23,689 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:28,690 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:33,545 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:50:33,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:50:34,652 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:39,652 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:44,653 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:48,371 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:50:48,546 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:50:48,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:50:49,714 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:54,715 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:50:59,716 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:03,546 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:51:03,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:51:05,639 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:10,640 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:15,640 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:18,373 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:51:18,546 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:51:18,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:51:20,719 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:25,719 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:30,720 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:33,546 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:51:33,546 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:51:36,677 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:41,809 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:51:42,008 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:45,922 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:51:47,894 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:48,375 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:51:48,835 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: stop_status +2024-05-29 13:51:48,835 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: stop_status +2024-05-29 13:51:52,972 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:54,180 DEBUG SenderThread:899 [sender.py:send():378] send: config +2024-05-29 13:51:54,181 DEBUG SenderThread:899 [sender.py:send():378] send: telemetry +2024-05-29 13:51:54,182 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: summary_record +2024-05-29 13:51:54,183 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: partial_history +2024-05-29 13:51:54,186 INFO SenderThread:899 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-29 13:51:54,186 DEBUG SenderThread:899 [sender.py:send():378] send: history +2024-05-29 13:51:54,187 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: summary_record +2024-05-29 13:51:54,188 INFO SenderThread:899 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-29 13:51:54,456 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: log_artifact +2024-05-29 13:51:54,457 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: log_artifact +2024-05-29 13:51:54,684 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: partial_history +2024-05-29 13:51:54,826 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: log_artifact +2024-05-29 13:51:54,939 INFO Thread-12 :899 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/wandb-summary.json +2024-05-29 13:51:54,939 INFO Thread-12 :899 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/media/table/evaluation/eval_results_1_f9e62f0af2531c926df9.table.json +2024-05-29 13:51:54,940 INFO Thread-12 :899 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/media/table/evaluation +2024-05-29 13:51:54,940 INFO Thread-12 :899 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/media +2024-05-29 13:51:54,940 INFO Thread-12 :899 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/media/table +2024-05-29 13:51:55,488 INFO wandb-upload_0:899 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpolvzyenf +2024-05-29 13:51:55,516 INFO wandb-upload_1:899 [upload_job.py:push():130] Uploaded file /tmp/tmp0s4o2ndwwandb/22qvx6sm-media/table/evaluation/eval_results_1_f9e62f0af2531c926df9.table.json +2024-05-29 13:51:55,936 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-29 13:51:55,948 INFO SenderThread:899 [sender.py:send_request_log_artifact():1455] logged artifact run-x16dn3vu-evaluationeval_results - {'id': 'QXJ0aWZhY3Q6ODUzNTI4Njk5', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE4MDIzNDMzOA==', 'latestArtifact': None}} +2024-05-29 13:51:55,949 DEBUG SenderThread:899 [sender.py:send():378] send: files +2024-05-29 13:51:55,949 INFO SenderThread:899 [sender.py:_save_file():1389] saving file media/table/evaluation/eval_results_1_f9e62f0af2531c926df9.table.json with policy now +2024-05-29 13:51:55,949 DEBUG SenderThread:899 [sender.py:send():378] send: history +2024-05-29 13:51:55,950 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: summary_record +2024-05-29 13:51:55,952 INFO SenderThread:899 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-29 13:51:55,953 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: log_artifact +2024-05-29 13:51:56,269 INFO wandb-upload_0:899 [upload_job.py:push():130] Uploaded file /tmp/tmp0s4o2ndwwandb/8iko1hd4-media/table/evaluation/eval_results_1_f9e62f0af2531c926df9.table.json +2024-05-29 13:51:56,641 INFO wandb-upload_1:899 [upload_job.py:push():88] Uploaded file /tmp/tmp98ty_bru/results.json +2024-05-29 13:51:56,943 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/wandb-summary.json +2024-05-29 13:51:57,136 INFO SenderThread:899 [sender.py:send_request_log_artifact():1455] logged artifact results - {'id': 'QXJ0aWZhY3Q6ODUzNTI4NzE3', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE4MDIyOTU4OQ==', 'latestArtifact': {'id': 'QXJ0aWZhY3Q6ODUzNTI4MjI0', 'versionIndex': 2}}} +2024-05-29 13:51:57,137 DEBUG SenderThread:899 [sender.py:send():378] send: telemetry +2024-05-29 13:51:57,137 DEBUG SenderThread:899 [sender.py:send():378] send: exit +2024-05-29 13:51:57,138 INFO SenderThread:899 [sender.py:send_exit():585] handling exit code: 0 +2024-05-29 13:51:57,138 INFO SenderThread:899 [sender.py:send_exit():587] handling runtime: 2766 +2024-05-29 13:51:57,139 INFO SenderThread:899 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-29 13:51:57,139 INFO SenderThread:899 [sender.py:send_exit():593] send defer +2024-05-29 13:51:57,139 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: poll_exit +2024-05-29 13:51:57,139 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:57,140 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-29 13:51:57,140 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:57,140 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-29 13:51:57,140 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 1 +2024-05-29 13:51:57,141 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:57,141 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-29 13:51:57,141 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:57,141 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-29 13:51:57,141 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 2 +2024-05-29 13:51:57,141 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:57,142 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-29 13:51:57,142 INFO HandlerThread:899 [system_monitor.py:finish():203] Stopping system monitor +2024-05-29 13:51:57,142 DEBUG SystemMonitor:899 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-29 13:51:57,143 DEBUG SystemMonitor:899 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-29 13:51:57,143 INFO HandlerThread:899 [interfaces.py:finish():200] Joined cpu monitor +2024-05-29 13:51:57,145 INFO HandlerThread:899 [interfaces.py:finish():200] Joined disk monitor +2024-05-29 13:51:57,145 INFO HandlerThread:899 [interfaces.py:finish():200] Joined memory monitor +2024-05-29 13:51:57,145 INFO HandlerThread:899 [interfaces.py:finish():200] Joined network monitor +2024-05-29 13:51:57,146 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:57,146 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-29 13:51:57,146 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 3 +2024-05-29 13:51:57,146 DEBUG SenderThread:899 [sender.py:send():378] send: stats +2024-05-29 13:51:57,146 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:57,148 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-29 13:51:57,148 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:57,148 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-29 13:51:57,148 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 4 +2024-05-29 13:51:57,148 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:57,148 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-29 13:51:57,149 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:57,149 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-29 13:51:57,149 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 5 +2024-05-29 13:51:57,149 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:57,149 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-29 13:51:57,150 DEBUG SenderThread:899 [sender.py:send():378] send: summary +2024-05-29 13:51:57,151 INFO SenderThread:899 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-29 13:51:57,151 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:57,153 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-29 13:51:57,154 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 6 +2024-05-29 13:51:57,154 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:57,154 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-29 13:51:57,154 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:57,154 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-29 13:51:57,158 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: status_report +2024-05-29 13:51:57,258 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 7 +2024-05-29 13:51:57,259 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:57,259 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-29 13:51:57,259 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:57,259 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-29 13:51:57,936 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-29 13:51:57,945 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/wandb-summary.json +2024-05-29 13:51:57,946 INFO Thread-12 :899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/config.yaml +2024-05-29 13:51:58,991 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 8 +2024-05-29 13:51:58,992 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: poll_exit +2024-05-29 13:51:58,992 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:58,992 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-29 13:51:58,993 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:58,993 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-29 13:51:58,993 INFO SenderThread:899 [job_builder.py:build():432] Attempting to build job artifact +2024-05-29 13:51:58,995 INFO SenderThread:899 [job_builder.py:_get_source_type():576] no source found +2024-05-29 13:51:58,995 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 9 +2024-05-29 13:51:58,995 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:58,995 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-29 13:51:58,996 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:58,996 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-29 13:51:58,996 INFO SenderThread:899 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-29 13:51:59,938 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-29 13:51:59,948 INFO SenderThread:899 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:51:59,948 INFO SenderThread:899 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files +2024-05-29 13:51:59,949 INFO SenderThread:899 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log output.log +2024-05-29 13:51:59,949 INFO SenderThread:899 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/requirements.txt requirements.txt +2024-05-29 13:51:59,951 INFO SenderThread:899 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/config.yaml config.yaml +2024-05-29 13:51:59,954 INFO SenderThread:899 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/wandb-summary.json wandb-summary.json +2024-05-29 13:51:59,954 INFO SenderThread:899 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/wandb-metadata.json wandb-metadata.json +2024-05-29 13:51:59,955 INFO SenderThread:899 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/media/table/evaluation/eval_results_1_f9e62f0af2531c926df9.table.json media/table/evaluation/eval_results_1_f9e62f0af2531c926df9.table.json +2024-05-29 13:51:59,955 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 10 +2024-05-29 13:51:59,955 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: poll_exit +2024-05-29 13:51:59,955 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:51:59,955 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-29 13:51:59,956 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:51:59,956 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-29 13:51:59,956 INFO SenderThread:899 [file_pusher.py:finish():169] shutting down file pusher +2024-05-29 13:52:00,205 INFO wandb-upload_0:899 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/output.log +2024-05-29 13:52:00,205 INFO wandb-upload_1:899 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/requirements.txt +2024-05-29 13:52:00,437 INFO wandb-upload_3:899 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/wandb-summary.json +2024-05-29 13:52:00,496 INFO wandb-upload_2:899 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/files/config.yaml +2024-05-29 13:52:00,697 INFO Thread-11 (_thread_body):899 [sender.py:transition_state():613] send defer: 11 +2024-05-29 13:52:00,697 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:52:00,697 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-29 13:52:00,698 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:52:00,698 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-29 13:52:00,698 INFO SenderThread:899 [file_pusher.py:join():175] waiting for file pusher +2024-05-29 13:52:00,699 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 12 +2024-05-29 13:52:00,699 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:52:00,699 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-29 13:52:00,699 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:52:00,702 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-29 13:52:00,702 INFO SenderThread:899 [file_stream.py:finish():601] file stream finish called +2024-05-29 13:52:00,938 INFO SenderThread:899 [file_stream.py:finish():605] file stream finish is done +2024-05-29 13:52:00,938 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 13 +2024-05-29 13:52:00,939 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:52:00,939 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-29 13:52:00,939 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-29 13:52:00,939 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:52:00,940 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-29 13:52:00,940 INFO SenderThread:899 [sender.py:transition_state():613] send defer: 14 +2024-05-29 13:52:00,940 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: poll_exit +2024-05-29 13:52:00,940 DEBUG SenderThread:899 [sender.py:send():378] send: final +2024-05-29 13:52:00,941 DEBUG SenderThread:899 [sender.py:send():378] send: footer +2024-05-29 13:52:00,940 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: defer +2024-05-29 13:52:00,941 INFO HandlerThread:899 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-29 13:52:00,941 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: defer +2024-05-29 13:52:00,941 INFO SenderThread:899 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-29 13:52:00,943 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-29 13:52:00,944 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: poll_exit +2024-05-29 13:52:00,945 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-29 13:52:00,946 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: server_info +2024-05-29 13:52:00,946 DEBUG SenderThread:899 [sender.py:send_request():405] send_request: server_info +2024-05-29 13:52:00,986 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: get_summary +2024-05-29 13:52:00,986 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-29 13:52:01,030 DEBUG HandlerThread:899 [handler.py:handle_request():158] handle_request: shutdown +2024-05-29 13:52:01,030 INFO HandlerThread:899 [handler.py:finish():882] shutting down handler +2024-05-29 13:52:01,946 INFO WriterThread:899 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/run-x16dn3vu.wandb +2024-05-29 13:52:02,010 INFO SenderThread:899 [sender.py:finish():1545] shutting down sender +2024-05-29 13:52:02,010 INFO SenderThread:899 [file_pusher.py:finish():169] shutting down file pusher +2024-05-29 13:52:02,010 INFO SenderThread:899 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/logs/debug.log b/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..53d5e72f17281dfbe5b12cae70bd8f549fd16052 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/logs/debug.log @@ -0,0 +1,36 @@ +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_setup.py:_flush():76] Configure stats pid to 743 +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-29 13:05:47,759 WARNING MainThread:743 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/logs/debug.log +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/logs/debug-internal.log +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_init.py:init():560] calling init triggers +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-29 13:05:47,759 INFO MainThread:743 [wandb_init.py:init():610] starting backend +2024-05-29 13:05:47,760 INFO MainThread:743 [wandb_init.py:init():614] setting up manager +2024-05-29 13:05:47,764 INFO MainThread:743 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-29 13:05:47,764 INFO MainThread:743 [wandb_init.py:init():622] backend started and connected +2024-05-29 13:05:47,768 INFO MainThread:743 [wandb_init.py:init():711] updated telemetry +2024-05-29 13:05:47,781 INFO MainThread:743 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-29 13:05:48,048 INFO MainThread:743 [wandb_run.py:_on_init():2396] communicating current version +2024-05-29 13:05:48,155 INFO MainThread:743 [wandb_run.py:_on_init():2405] got version response +2024-05-29 13:05:48,156 INFO MainThread:743 [wandb_init.py:init():795] starting run threads in backend +2024-05-29 13:05:48,437 INFO MainThread:743 [wandb_run.py:_console_start():2374] atexit reg +2024-05-29 13:05:48,437 INFO MainThread:743 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-29 13:05:48,437 INFO MainThread:743 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-29 13:05:48,437 INFO MainThread:743 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-29 13:05:48,439 INFO MainThread:743 [wandb_init.py:init():838] run started, returning control to user process +2024-05-29 13:51:54,178 INFO MainThread:743 [wandb_run.py:_config_callback():1376] config_cb None None {'task_configs': {'arc_easy': {'task': 'arc_easy', 'group': ['ai2_arc'], 'dataset_path': 'allenai/ai2_arc', 'dataset_name': 'ARC-Easy', 'training_split': 'train', 'validation_split': 'validation', 'test_split': 'test', 'doc_to_text': 'Question: {{question}}\nAnswer:', 'doc_to_target': '{{choices.label.index(answerKey)}}', 'doc_to_choice': '{{choices.text}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}, {'metric': 'acc_norm', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'Question: {{question}}\nAnswer:', 'metadata': {'version': 1.0}}, 'boolq': {'task': 'boolq', 'group': ['super-glue-lm-eval-v1'], 'dataset_path': 'super_glue', 'dataset_name': 'boolq', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': '{{passage}}\nQuestion: {{question}}?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['no', 'yes'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'passage', 'metadata': {'version': 2.0}}, 'copa': {'task': 'copa', 'group': ['super-glue-lm-eval-v1'], 'dataset_path': 'super_glue', 'dataset_name': 'copa', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'def doc_to_text(doc):\n # Drop the period\n connector = {\n "cause": "because",\n "effect": "therefore",\n }[doc["question"]]\n return doc["premise"].strip()[:-1] + f" {connector}"\n', 'doc_to_target': 'def doc_to_target(doc):\n correct_choice = doc["choice1"] if doc["label"] == 0 else doc["choice2"]\n # Connect the sentences\n return " " + convert_choice(correct_choice)\n', 'doc_to_choice': 'def doc_to_choice(doc):\n return [" " + convert_choice(doc["choice1"]), " " + convert_choice(doc["choice2"])]\n', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'mrpc': {'task': 'mrpc', 'group': 'glue', 'dataset_path': 'glue', 'dataset_name': 'mrpc', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'Sentence 1: {{sentence1}}\nSentence 2: {{sentence2}}\nQuestion: Do both sentences mean the same thing?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['no', 'yes'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}, {'metric': 'f1'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'piqa': {'task': 'piqa', 'dataset_path': 'piqa', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'Question: {{goal}}\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': '{{[sol1, sol2]}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}, {'metric': 'acc_norm', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'goal', 'metadata': {'version': 1.0}}, 'sst2': {'task': 'sst2', 'group': 'glue', 'dataset_path': 'glue', 'dataset_name': 'sst2', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': '{{sentence}}\nQuestion: Is this sentence positive or negative?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['negative', 'positive'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'winogrande': {'task': 'winogrande', 'dataset_path': 'winogrande', 'dataset_name': 'winogrande_xl', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'def doc_to_text(doc):\n answer_to_num = {"1": 0, "2": 1}\n return answer_to_num[doc["answer"]]\n', 'doc_to_target': 'def doc_to_target(doc):\n idx = doc["sentence"].index("_") + 1\n return doc["sentence"][idx:].strip()\n', 'doc_to_choice': 'def doc_to_choice(doc):\n idx = doc["sentence"].index("_")\n options = [doc["option1"], doc["option2"]]\n return [doc["sentence"][:idx] + opt for opt in options]\n', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'sentence', 'metadata': {'version': 1.0}}}, 'cli_configs': {'model': 'hf', 'model_args': 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint/llamav2-3b/hf/global_step40000,tokenizer=huggyllama/llama-13b', 'batch_size': 'auto', 'batch_sizes': [64], 'device': None, 'use_cache': None, 'limit': None, 'bootstrap_iters': 100000, 'gen_kwargs': None}} +2024-05-29 13:51:54,934 INFO MainThread:743 [wandb_run.py:_finish():2103] finishing run smlgenai/english-eval/x16dn3vu +2024-05-29 13:51:54,934 INFO MainThread:743 [wandb_run.py:_atexit_cleanup():2343] got exitcode: 0 +2024-05-29 13:51:54,935 INFO MainThread:743 [wandb_run.py:_restore():2326] restore +2024-05-29 13:51:54,935 INFO MainThread:743 [wandb_run.py:_restore():2332] restore done +2024-05-29 13:52:02,032 INFO MainThread:743 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-29 13:52:02,032 INFO MainThread:743 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-29 13:52:02,039 INFO MainThread:743 [wandb_run.py:_footer_sync_info():3953] logging synced files diff --git a/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/run-x16dn3vu.wandb b/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/run-x16dn3vu.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ad111c4ece4d6536541c6a035aee9f9965fcd7ad Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240529_130547-x16dn3vu/run-x16dn3vu.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/config.yaml b/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..dcff9677d6c04e2cca35a1ec54c325ba7da71450 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/config.yaml @@ -0,0 +1,283 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.36.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716988011 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 2 + - 23 + - 62 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.36.2 + 8: + - 5 + 13: linux-x86_64 +task_configs: + desc: null + value: + arc_easy: + task: arc_easy + group: + - ai2_arc + dataset_path: allenai/ai2_arc + dataset_name: ARC-Easy + training_split: train + validation_split: validation + test_split: test + doc_to_text: 'Question: {{question}} + + Answer:' + doc_to_target: '{{choices.label.index(answerKey)}}' + doc_to_choice: '{{choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{question}} + + Answer:' + metadata: + version: 1.0 + boolq: + task: boolq + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: boolq + training_split: train + validation_split: validation + doc_to_text: '{{passage}} + + Question: {{question}}? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: passage + metadata: + version: 2.0 + copa: + task: copa + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: copa + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n # Drop the period\n connector =\ + \ {\n \"cause\": \"because\",\n \"effect\": \"therefore\",\n\ + \ }[doc[\"question\"]]\n return doc[\"premise\"].strip()[:-1] + f\"\ + \ {connector}\"\n" + doc_to_target: "def doc_to_target(doc):\n correct_choice = doc[\"choice1\"\ + ] if doc[\"label\"] == 0 else doc[\"choice2\"]\n # Connect the sentences\n\ + \ return \" \" + convert_choice(correct_choice)\n" + doc_to_choice: "def doc_to_choice(doc):\n return [\" \" + convert_choice(doc[\"\ + choice1\"]), \" \" + convert_choice(doc[\"choice2\"])]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + mrpc: + task: mrpc + group: glue + dataset_path: glue + dataset_name: mrpc + training_split: train + validation_split: validation + doc_to_text: 'Sentence 1: {{sentence1}} + + Sentence 2: {{sentence2}} + + Question: Do both sentences mean the same thing? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + - metric: f1 + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + piqa: + task: piqa + dataset_path: piqa + training_split: train + validation_split: validation + doc_to_text: 'Question: {{goal}} + + Answer:' + doc_to_target: label + doc_to_choice: '{{[sol1, sol2]}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: goal + metadata: + version: 1.0 + sst2: + task: sst2 + group: glue + dataset_path: glue + dataset_name: sst2 + training_split: train + validation_split: validation + doc_to_text: '{{sentence}} + + Question: Is this sentence positive or negative? + + Answer:' + doc_to_target: label + doc_to_choice: + - negative + - positive + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + winogrande: + task: winogrande + dataset_path: winogrande + dataset_name: winogrande_xl + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n answer_to_num = {\"1\": 0, \"2\": 1}\n\ + \ return answer_to_num[doc[\"answer\"]]\n" + doc_to_target: "def doc_to_target(doc):\n idx = doc[\"sentence\"].index(\"\ + _\") + 1\n return doc[\"sentence\"][idx:].strip()\n" + doc_to_choice: "def doc_to_choice(doc):\n idx = doc[\"sentence\"].index(\"\ + _\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"\ + sentence\"][:idx] + opt for opt in options]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: sentence + metadata: + version: 1.0 +cli_configs: + desc: null + value: + model: hf + model_args: pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint/llamav2-3b/hf/global_step70000,tokenizer=huggyllama/llama-13b + batch_size: auto + batch_sizes: + - 64 + device: null + use_cache: null + limit: null + bootstrap_iters: 100000 + gen_kwargs: null diff --git a/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/media/table/evaluation/eval_results_1_3da77a80c594c9e07db8.table.json b/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/media/table/evaluation/eval_results_1_3da77a80c594c9e07db8.table.json new file mode 100644 index 0000000000000000000000000000000000000000..d1a06fa77734fed15bd0993a96331531a6ca6f9d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/media/table/evaluation/eval_results_1_3da77a80c594c9e07db8.table.json @@ -0,0 +1 @@ +{"columns": ["Tasks", "Version", "Filter", "num_fewshot", "Metric", "Value", "Stderr"], "data": [["winogrande", 1.0, "none", 0, "acc", "0.4956590370955012", "0.0141"], ["sst2", 1.0, "none", 0, "acc", "0.5", "0.0169"], ["piqa", 1.0, "none", 0, "acc", "0.5087051142546246", "0.0117"], ["piqa", 1.0, "none", 0, "acc_norm", "0.4967355821545158", "0.0117"], ["mrpc", 1.0, "none", 0, "acc", "0.5245098039215687", "0.0248"], ["mrpc", 1.0, "none", 0, "f1", "0.6269230769230769", "0.0248"], ["copa", 1.0, "none", 0, "acc", "0.61", "0.0490"], ["boolq", 2.0, "none", 0, "acc", "0.491131498470948", "0.0087"], ["arc_easy", 1.0, "none", 0, "acc", "0.2638888888888889", "0.0090"], ["arc_easy", 1.0, "none", 0, "acc_norm", "0.25841750841750843", "0.0090"]]} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/output.log b/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e42ad69f281aafc0c3f34bd1a50c66637f9ad9a8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240529_130651-temx2bmb/files/output.log @@ -0,0 +1,479 @@ + +2024-05-29:13:06:52,082 INFO [__main__.py:251] Verbosity set to INFO +2024-05-29:13:07:01,670 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'boolq', 'copa', 'mrpc', 'piqa', 'sst2', 'winogrande'] +2024-05-29:13:07:01,672 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-29:13:07:01,672 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/eval/checkpoint/llamav2-3b/hf/global_step70000', 'tokenizer': 'huggyllama/llama-13b'} +2024-05-29:13:07:03,957 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`. + warnings.warn( +Downloading readme: 100%|██████████| 9.00k/9.00k [00:00<00:00, 16.5MB/s] +Downloading data: 100%|██████████| 331k/331k [00:00<00:00, 3.51MB/s] +Downloading data: 100%|██████████| 346k/346k [00:00<00:00, 4.54MB/s] +Downloading data: 100%|██████████| 86.1k/86.1k [00:00<00:00, 1.17MB/s] +Generating train split: 100%|██████████| 2251/2251 [00:00<00:00, 96969.91 examples/s] +Generating test split: 100%|██████████| 2376/2376 [00:00<00:00, 331868.07 examples/s] +Generating validation split: 100%|██████████| 570/570 [00:00<00:00, 191582.12 examples/s] +2024-05-29:13:07:34,137 WARNING [task.py:763] [Task: boolq] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-29:13:07:34,138 WARNING [task.py:775] [Task: boolq] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for super_glue contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/super_glue +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 30.7k/30.7k [00:00<00:00, 40.1MB/s] +Downloading readme: 100%|██████████| 18.2k/18.2k [00:00<00:00, 27.6MB/s] +Downloading data: 100%|██████████| 4.12M/4.12M [00:00<00:00, 75.1MB/s] +Generating train split: 100%|██████████| 9427/9427 [00:00<00:00, 22556.97 examples/s] +Generating validation split: 100%|██████████| 3270/3270 [00:00<00:00, 21982.65 examples/s] +Generating test split: 100%|██████████| 3245/3245 [00:00<00:00, 23389.95 examples/s] +2024-05-29:13:07:42,071 WARNING [task.py:763] [Task: copa] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-29:13:07:42,072 WARNING [task.py:775] [Task: copa] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +Downloading data: 100%|██████████| 44.0k/44.0k [00:00<00:00, 46.3MB/s] +Generating train split: 100%|██████████| 400/400 [00:00<00:00, 16362.75 examples/s] +Generating validation split: 100%|██████████| 100/100 [00:00<00:00, 12810.56 examples/s] +Generating test split: 100%|██████████| 500/500 [00:00<00:00, 17401.58 examples/s] +2024-05-29:13:07:44,824 WARNING [task.py:763] [Task: mrpc] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-29:13:07:44,825 WARNING [task.py:775] [Task: mrpc] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +2024-05-29:13:07:44,826 WARNING [task.py:763] [Task: mrpc] metric f1 is defined, but aggregation is not. using default aggregation=f1 +2024-05-29:13:07:44,826 WARNING [task.py:775] [Task: mrpc] metric f1 is defined, but higher_is_better is not. using default higher_is_better=True +Downloading readme: 100%|██████████| 35.3k/35.3k [00:00<00:00, 643kB/s] +Downloading data: 100%|██████████| 649k/649k [00:00<00:00, 4.00MB/s] +Downloading data: 100%|██████████| 75.7k/75.7k [00:00<00:00, 529kB/s] +Downloading data: 100%|██████████| 308k/308k [00:00<00:00, 2.06MB/s] +Generating train split: 100%|██████████| 3668/3668 [00:00<00:00, 408407.41 examples/s] +Generating validation split: 100%|██████████| 408/408 [00:00<00:00, 174709.14 examples/s] +Generating test split: 100%|██████████| 1725/1725 [00:00<00:00, 364492.41 examples/s] +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for piqa contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/piqa +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 5.36k/5.36k [00:00<00:00, 11.8MB/s] +Downloading readme: 100%|██████████| 8.41k/8.41k [00:00<00:00, 17.5MB/s] +Downloading data: 100%|██████████| 1.82M/1.82M [00:00<00:00, 4.22MB/s] +Downloading data: 100%|██████████| 815k/815k [00:00<00:00, 22.1MB/s] +Generating train split: 100%|██████████| 16113/16113 [00:00<00:00, 24127.79 examples/s] +Generating test split: 100%|██████████| 3084/3084 [00:00<00:00, 24081.37 examples/s] +Generating validation split: 100%|██████████| 1838/1838 [00:00<00:00, 24195.38 examples/s] +2024-05-29:13:07:55,745 WARNING [task.py:763] [Task: sst2] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-29:13:07:55,746 WARNING [task.py:775] [Task: sst2] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +Downloading data: 100%|██████████| 3.11M/3.11M [00:00<00:00, 20.4MB/s] +Downloading data: 100%|██████████| 72.8k/72.8k [00:00<00:00, 502kB/s] +Downloading data: 100%|██████████| 148k/148k [00:00<00:00, 993kB/s] +Generating train split: 100%|██████████| 67349/67349 [00:00<00:00, 1428582.45 examples/s] +Generating validation split: 100%|██████████| 872/872 [00:00<00:00, 398500.01 examples/s] +Generating test split: 100%|██████████| 1821/1821 [00:00<00:00, 533815.18 examples/s] +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for winogrande contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/winogrande +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 5.65k/5.65k [00:00<00:00, 11.5MB/s] +Downloading readme: 100%|██████████| 9.97k/9.97k [00:00<00:00, 20.3MB/s] +Downloading data: 100%|██████████| 3.40M/3.40M [00:00<00:00, 6.93MB/s] +Generating train split: 100%|██████████| 40398/40398 [00:01<00:00, 24570.02 examples/s] +Generating test split: 100%|██████████| 1767/1767 [00:00<00:00, 24632.20 examples/s] +Generating validation split: 100%|██████████| 1267/1267 [00:00<00:00, 24135.85 examples/s] +2024-05-29:13:08:08,455 INFO [task.py:395] Building contexts for winogrande on rank 0... +100%|██████████| 1267/1267 [00:00<00:00, 69899.55it/s] +2024-05-29:13:08:08,539 INFO [task.py:395] Building contexts for sst2 on rank 0... +100%|██████████| 872/872 [00:00<00:00, 2559.22it/s] +2024-05-29:13:08:08,910 INFO [task.py:395] Building contexts for piqa on rank 0... +100%|██████████| 1838/1838 [00:01<00:00, 1085.76it/s] +2024-05-29:13:08:10,676 INFO [task.py:395] Building contexts for mrpc on rank 0... +100%|██████████| 408/408 [00:00<00:00, 1839.14it/s] +2024-05-29:13:08:10,917 INFO [task.py:395] Building contexts for copa on rank 0... +100%|██████████| 100/100 [00:00<00:00, 61052.46it/s] +2024-05-29:13:08:10,926 INFO [task.py:395] Building contexts for boolq on rank 0... +100%|██████████| 3270/3270 [00:01<00:00, 1986.01it/s] +2024-05-29:13:08:12,703 INFO [task.py:395] Building contexts for arc_easy on rank 0... +100%|██████████| 2376/2376 [00:02<00:00, 1072.52it/s] +2024-05-29:13:08:15,064 INFO [evaluator.py:379] Running loglikelihood requests +Running loglikelihood requests: 0%| | 0/25011 [00:00=2.1.1. The default is otherwise the manual `"eager"` implementation. + + Examples: + + ```python + >>> from transformers import AutoConfig, BaseAutoModelClass + + >>> # Download configuration from huggingface.co and cache. + >>> config = AutoConfig.from_pretrained("checkpoint_placeholder") + >>> model = BaseAutoModelClass.from_config(config) + ``` +""" + +FROM_PRETRAINED_TORCH_DOCSTRING = """ + Instantiate one of the model classes of the library from a pretrained model. + + The model class to instantiate is selected based on the `model_type` property of the config object (either + passed as an argument or loaded from `pretrained_model_name_or_path` if possible), or when it's missing, by + falling back to using pattern matching on `pretrained_model_name_or_path`: + + List options + + The model is set in evaluation mode by default using `model.eval()` (so for instance, dropout modules are + deactivated). To train the model, you should first set it back in training mode with `model.train()` + + Args: + pretrained_model_name_or_path (`str` or `os.PathLike`): + Can be either: + + - A string, the *model id* of a pretrained model hosted inside a model repo on huggingface.co. + - A path to a *directory* containing model weights saved using + [`~PreTrainedModel.save_pretrained`], e.g., `./my_model_directory/`. + - A path or url to a *tensorflow index checkpoint file* (e.g, `./tf_model/model.ckpt.index`). In + this case, `from_tf` should be set to `True` and a configuration object should be provided as + `config` argument. This loading path is slower than converting the TensorFlow checkpoint in a + PyTorch model using the provided conversion scripts and loading the PyTorch model afterwards. + model_args (additional positional arguments, *optional*): + Will be passed along to the underlying model `__init__()` method. + config ([`PretrainedConfig`], *optional*): + Configuration for the model to use instead of an automatically loaded configuration. Configuration can + be automatically loaded when: + + - The model is a model provided by the library (loaded with the *model id* string of a pretrained + model). + - The model was saved using [`~PreTrainedModel.save_pretrained`] and is reloaded by supplying the + save directory. + - The model is loaded by supplying a local directory as `pretrained_model_name_or_path` and a + configuration JSON file named *config.json* is found in the directory. + state_dict (*Dict[str, torch.Tensor]*, *optional*): + A state dictionary to use instead of a state dictionary loaded from saved weights file. + + This option can be used if you want to create a model from a pretrained configuration but load your own + weights. In this case though, you should check if using [`~PreTrainedModel.save_pretrained`] and + [`~PreTrainedModel.from_pretrained`] is not a simpler option. + cache_dir (`str` or `os.PathLike`, *optional*): + Path to a directory in which a downloaded pretrained model configuration should be cached if the + standard cache should not be used. + from_tf (`bool`, *optional*, defaults to `False`): + Load the model weights from a TensorFlow checkpoint save file (see docstring of + `pretrained_model_name_or_path` argument). + force_download (`bool`, *optional*, defaults to `False`): + Whether or not to force the (re-)download of the model weights and configuration files, overriding the + cached versions if they exist. + resume_download (`bool`, *optional*, defaults to `False`): + Whether or not to delete incompletely received files. Will attempt to resume the download if such a + file exists. + proxies (`Dict[str, str]`, *optional*): + A dictionary of proxy servers to use by protocol or endpoint, e.g., `{'http': 'foo.bar:3128', + 'http://hostname': 'foo.bar:4012'}`. The proxies are used on each request. + output_loading_info(`bool`, *optional*, defaults to `False`): + Whether ot not to also return a dictionary containing missing keys, unexpected keys and error messages. + local_files_only(`bool`, *optional*, defaults to `False`): + Whether or not to only look at local files (e.g., not try downloading the model). + revision (`str`, *optional*, defaults to `"main"`): + The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a + git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any + identifier allowed by git. + trust_remote_code (`bool`, *optional*, defaults to `False`): + Whether or not to allow for custom models defined on the Hub in their own modeling files. This option + should only be set to `True` for repositories you trust and in which you have read the code, as it will + execute code present on the Hub on your local machine. + code_revision (`str`, *optional*, defaults to `"main"`): + The specific revision to use for the code on the Hub, if the code leaves in a different repository than + the rest of the model. It can be a branch name, a tag name, or a commit id, since we use a git-based + system for storing models and other artifacts on huggingface.co, so `revision` can be any identifier + allowed by git. + kwargs (additional keyword arguments, *optional*): + Can be used to update the configuration object (after it being loaded) and initiate the model (e.g., + `output_attentions=True`). Behaves differently depending on whether a `config` is provided or + automatically loaded: + + - If a configuration is provided with `config`, `**kwargs` will be directly passed to the + underlying model's `__init__` method (we assume all relevant updates to the configuration have + already been done) + - If a configuration is not provided, `kwargs` will be first passed to the configuration class + initialization function ([`~PretrainedConfig.from_pretrained`]). Each key of `kwargs` that + corresponds to a configuration attribute will be used to override said attribute with the + supplied `kwargs` value. Remaining keys that do not correspond to any configuration attribute + will be passed to the underlying model's `__init__` function. + + Examples: + + ```python + >>> from transformers import AutoConfig, BaseAutoModelClass + + >>> # Download model and configuration from huggingface.co and cache. + >>> model = BaseAutoModelClass.from_pretrained("checkpoint_placeholder") + + >>> # Update configuration during loading + >>> model = BaseAutoModelClass.from_pretrained("checkpoint_placeholder", output_attentions=True) + >>> model.config.output_attentions + True + + >>> # Loading from a TF checkpoint file instead of a PyTorch model (slower) + >>> config = AutoConfig.from_pretrained("./tf_model/shortcut_placeholder_tf_model_config.json") + >>> model = BaseAutoModelClass.from_pretrained( + ... "./tf_model/shortcut_placeholder_tf_checkpoint.ckpt.index", from_tf=True, config=config + ... ) + ``` +""" + +FROM_PRETRAINED_TF_DOCSTRING = """ + Instantiate one of the model classes of the library from a pretrained model. + + The model class to instantiate is selected based on the `model_type` property of the config object (either + passed as an argument or loaded from `pretrained_model_name_or_path` if possible), or when it's missing, by + falling back to using pattern matching on `pretrained_model_name_or_path`: + + List options + + Args: + pretrained_model_name_or_path (`str` or `os.PathLike`): + Can be either: + + - A string, the *model id* of a pretrained model hosted inside a model repo on huggingface.co. + - A path to a *directory* containing model weights saved using + [`~PreTrainedModel.save_pretrained`], e.g., `./my_model_directory/`. + - A path or url to a *PyTorch state_dict save file* (e.g, `./pt_model/pytorch_model.bin`). In this + case, `from_pt` should be set to `True` and a configuration object should be provided as `config` + argument. This loading path is slower than converting the PyTorch model in a TensorFlow model + using the provided conversion scripts and loading the TensorFlow model afterwards. + model_args (additional positional arguments, *optional*): + Will be passed along to the underlying model `__init__()` method. + config ([`PretrainedConfig`], *optional*): + Configuration for the model to use instead of an automatically loaded configuration. Configuration can + be automatically loaded when: + + - The model is a model provided by the library (loaded with the *model id* string of a pretrained + model). + - The model was saved using [`~PreTrainedModel.save_pretrained`] and is reloaded by supplying the + save directory. + - The model is loaded by supplying a local directory as `pretrained_model_name_or_path` and a + configuration JSON file named *config.json* is found in the directory. + cache_dir (`str` or `os.PathLike`, *optional*): + Path to a directory in which a downloaded pretrained model configuration should be cached if the + standard cache should not be used. + from_pt (`bool`, *optional*, defaults to `False`): + Load the model weights from a PyTorch checkpoint save file (see docstring of + `pretrained_model_name_or_path` argument). + force_download (`bool`, *optional*, defaults to `False`): + Whether or not to force the (re-)download of the model weights and configuration files, overriding the + cached versions if they exist. + resume_download (`bool`, *optional*, defaults to `False`): + Whether or not to delete incompletely received files. Will attempt to resume the download if such a + file exists. + proxies (`Dict[str, str]`, *optional*): + A dictionary of proxy servers to use by protocol or endpoint, e.g., `{'http': 'foo.bar:3128', + 'http://hostname': 'foo.bar:4012'}`. The proxies are used on each request. + output_loading_info(`bool`, *optional*, defaults to `False`): + Whether ot not to also return a dictionary containing missing keys, unexpected keys and error messages. + local_files_only(`bool`, *optional*, defaults to `False`): + Whether or not to only look at local files (e.g., not try downloading the model). + revision (`str`, *optional*, defaults to `"main"`): + The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a + git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any + identifier allowed by git. + trust_remote_code (`bool`, *optional*, defaults to `False`): + Whether or not to allow for custom models defined on the Hub in their own modeling files. This option + should only be set to `True` for repositories you trust and in which you have read the code, as it will + execute code present on the Hub on your local machine. + code_revision (`str`, *optional*, defaults to `"main"`): + The specific revision to use for the code on the Hub, if the code leaves in a different repository than + the rest of the model. It can be a branch name, a tag name, or a commit id, since we use a git-based + system for storing models and other artifacts on huggingface.co, so `revision` can be any identifier + allowed by git. + kwargs (additional keyword arguments, *optional*): + Can be used to update the configuration object (after it being loaded) and initiate the model (e.g., + `output_attentions=True`). Behaves differently depending on whether a `config` is provided or + automatically loaded: + + - If a configuration is provided with `config`, `**kwargs` will be directly passed to the + underlying model's `__init__` method (we assume all relevant updates to the configuration have + already been done) + - If a configuration is not provided, `kwargs` will be first passed to the configuration class + initialization function ([`~PretrainedConfig.from_pretrained`]). Each key of `kwargs` that + corresponds to a configuration attribute will be used to override said attribute with the + supplied `kwargs` value. Remaining keys that do not correspond to any configuration attribute + will be passed to the underlying model's `__init__` function. + + Examples: + + ```python + >>> from transformers import AutoConfig, BaseAutoModelClass + + >>> # Download model and configuration from huggingface.co and cache. + >>> model = BaseAutoModelClass.from_pretrained("checkpoint_placeholder") + + >>> # Update configuration during loading + >>> model = BaseAutoModelClass.from_pretrained("checkpoint_placeholder", output_attentions=True) + >>> model.config.output_attentions + True + + >>> # Loading from a PyTorch checkpoint file instead of a TensorFlow model (slower) + >>> config = AutoConfig.from_pretrained("./pt_model/shortcut_placeholder_pt_model_config.json") + >>> model = BaseAutoModelClass.from_pretrained( + ... "./pt_model/shortcut_placeholder_pytorch_model.bin", from_pt=True, config=config + ... ) + ``` +""" + +FROM_PRETRAINED_FLAX_DOCSTRING = """ + Instantiate one of the model classes of the library from a pretrained model. + + The model class to instantiate is selected based on the `model_type` property of the config object (either + passed as an argument or loaded from `pretrained_model_name_or_path` if possible), or when it's missing, by + falling back to using pattern matching on `pretrained_model_name_or_path`: + + List options + + Args: + pretrained_model_name_or_path (`str` or `os.PathLike`): + Can be either: + + - A string, the *model id* of a pretrained model hosted inside a model repo on huggingface.co. + - A path to a *directory* containing model weights saved using + [`~PreTrainedModel.save_pretrained`], e.g., `./my_model_directory/`. + - A path or url to a *PyTorch state_dict save file* (e.g, `./pt_model/pytorch_model.bin`). In this + case, `from_pt` should be set to `True` and a configuration object should be provided as `config` + argument. This loading path is slower than converting the PyTorch model in a TensorFlow model + using the provided conversion scripts and loading the TensorFlow model afterwards. + model_args (additional positional arguments, *optional*): + Will be passed along to the underlying model `__init__()` method. + config ([`PretrainedConfig`], *optional*): + Configuration for the model to use instead of an automatically loaded configuration. Configuration can + be automatically loaded when: + + - The model is a model provided by the library (loaded with the *model id* string of a pretrained + model). + - The model was saved using [`~PreTrainedModel.save_pretrained`] and is reloaded by supplying the + save directory. + - The model is loaded by supplying a local directory as `pretrained_model_name_or_path` and a + configuration JSON file named *config.json* is found in the directory. + cache_dir (`str` or `os.PathLike`, *optional*): + Path to a directory in which a downloaded pretrained model configuration should be cached if the + standard cache should not be used. + from_pt (`bool`, *optional*, defaults to `False`): + Load the model weights from a PyTorch checkpoint save file (see docstring of + `pretrained_model_name_or_path` argument). + force_download (`bool`, *optional*, defaults to `False`): + Whether or not to force the (re-)download of the model weights and configuration files, overriding the + cached versions if they exist. + resume_download (`bool`, *optional*, defaults to `False`): + Whether or not to delete incompletely received files. Will attempt to resume the download if such a + file exists. + proxies (`Dict[str, str]`, *optional*): + A dictionary of proxy servers to use by protocol or endpoint, e.g., `{'http': 'foo.bar:3128', + 'http://hostname': 'foo.bar:4012'}`. The proxies are used on each request. + output_loading_info(`bool`, *optional*, defaults to `False`): + Whether ot not to also return a dictionary containing missing keys, unexpected keys and error messages. + local_files_only(`bool`, *optional*, defaults to `False`): + Whether or not to only look at local files (e.g., not try downloading the model). + revision (`str`, *optional*, defaults to `"main"`): + The specific model version to use. It can be a branch name, a tag name, or a commit id, since we use a + git-based system for storing models and other artifacts on huggingface.co, so `revision` can be any + identifier allowed by git. + trust_remote_code (`bool`, *optional*, defaults to `False`): + Whether or not to allow for custom models defined on the Hub in their own modeling files. This option + should only be set to `True` for repositories you trust and in which you have read the code, as it will + execute code present on the Hub on your local machine. + code_revision (`str`, *optional*, defaults to `"main"`): + The specific revision to use for the code on the Hub, if the code leaves in a different repository than + the rest of the model. It can be a branch name, a tag name, or a commit id, since we use a git-based + system for storing models and other artifacts on huggingface.co, so `revision` can be any identifier + allowed by git. + kwargs (additional keyword arguments, *optional*): + Can be used to update the configuration object (after it being loaded) and initiate the model (e.g., + `output_attentions=True`). Behaves differently depending on whether a `config` is provided or + automatically loaded: + + - If a configuration is provided with `config`, `**kwargs` will be directly passed to the + underlying model's `__init__` method (we assume all relevant updates to the configuration have + already been done) + - If a configuration is not provided, `kwargs` will be first passed to the configuration class + initialization function ([`~PretrainedConfig.from_pretrained`]). Each key of `kwargs` that + corresponds to a configuration attribute will be used to override said attribute with the + supplied `kwargs` value. Remaining keys that do not correspond to any configuration attribute + will be passed to the underlying model's `__init__` function. + + Examples: + + ```python + >>> from transformers import AutoConfig, BaseAutoModelClass + + >>> # Download model and configuration from huggingface.co and cache. + >>> model = BaseAutoModelClass.from_pretrained("checkpoint_placeholder") + + >>> # Update configuration during loading + >>> model = BaseAutoModelClass.from_pretrained("checkpoint_placeholder", output_attentions=True) + >>> model.config.output_attentions + True + + >>> # Loading from a PyTorch checkpoint file instead of a TensorFlow model (slower) + >>> config = AutoConfig.from_pretrained("./pt_model/shortcut_placeholder_pt_model_config.json") + >>> model = BaseAutoModelClass.from_pretrained( + ... "./pt_model/shortcut_placeholder_pytorch_model.bin", from_pt=True, config=config + ... ) + ``` +""" + + +def _get_model_class(config, model_mapping): + supported_models = model_mapping[type(config)] + if not isinstance(supported_models, (list, tuple)): + return supported_models + + name_to_model = {model.__name__: model for model in supported_models} + architectures = getattr(config, "architectures", []) + for arch in architectures: + if arch in name_to_model: + return name_to_model[arch] + elif f"TF{arch}" in name_to_model: + return name_to_model[f"TF{arch}"] + elif f"Flax{arch}" in name_to_model: + return name_to_model[f"Flax{arch}"] + + # If not architecture is set in the config or match the supported models, the first element of the tuple is the + # defaults. + return supported_models[0] + + +class _BaseAutoModelClass: + # Base class for auto models. + _model_mapping = None + + def __init__(self, *args, **kwargs): + raise EnvironmentError( + f"{self.__class__.__name__} is designed to be instantiated " + f"using the `{self.__class__.__name__}.from_pretrained(pretrained_model_name_or_path)` or " + f"`{self.__class__.__name__}.from_config(config)` methods." + ) + + @classmethod + def from_config(cls, config, **kwargs): + trust_remote_code = kwargs.pop("trust_remote_code", None) + has_remote_code = hasattr(config, "auto_map") and cls.__name__ in config.auto_map + has_local_code = type(config) in cls._model_mapping.keys() + trust_remote_code = resolve_trust_remote_code( + trust_remote_code, config._name_or_path, has_local_code, has_remote_code + ) + + if has_remote_code and trust_remote_code: + class_ref = config.auto_map[cls.__name__] + if "--" in class_ref: + repo_id, class_ref = class_ref.split("--") + else: + repo_id = config.name_or_path + model_class = get_class_from_dynamic_module(class_ref, repo_id, **kwargs) + if os.path.isdir(config._name_or_path): + model_class.register_for_auto_class(cls.__name__) + else: + cls.register(config.__class__, model_class, exist_ok=True) + _ = kwargs.pop("code_revision", None) + return model_class._from_config(config, **kwargs) + elif type(config) in cls._model_mapping.keys(): + model_class = _get_model_class(config, cls._model_mapping) + return model_class._from_config(config, **kwargs) + + raise ValueError( + f"Unrecognized configuration class {config.__class__} for this kind of AutoModel: {cls.__name__}.\n" + f"Model type should be one of {', '.join(c.__name__ for c in cls._model_mapping.keys())}." + ) + + @classmethod + def from_pretrained(cls, pretrained_model_name_or_path, *model_args, **kwargs): + config = kwargs.pop("config", None) + trust_remote_code = kwargs.pop("trust_remote_code", None) + kwargs["_from_auto"] = True + hub_kwargs_names = [ + "cache_dir", + "force_download", + "local_files_only", + "proxies", + "resume_download", + "revision", + "subfolder", + "use_auth_token", + "token", + ] + hub_kwargs = {name: kwargs.pop(name) for name in hub_kwargs_names if name in kwargs} + code_revision = kwargs.pop("code_revision", None) + commit_hash = kwargs.pop("_commit_hash", None) + adapter_kwargs = kwargs.pop("adapter_kwargs", None) + + token = hub_kwargs.pop("token", None) + use_auth_token = hub_kwargs.pop("use_auth_token", None) + if use_auth_token is not None: + warnings.warn( + "The `use_auth_token` argument is deprecated and will be removed in v5 of Transformers. Please use `token` instead.", + FutureWarning, + ) + if token is not None: + raise ValueError( + "`token` and `use_auth_token` are both specified. Please set only the argument `token`." + ) + token = use_auth_token + + if token is not None: + hub_kwargs["token"] = token + + if commit_hash is None: + if not isinstance(config, PretrainedConfig): + # We make a call to the config file first (which may be absent) to get the commit hash as soon as possible + resolved_config_file = cached_file( + pretrained_model_name_or_path, + CONFIG_NAME, + _raise_exceptions_for_gated_repo=False, + _raise_exceptions_for_missing_entries=False, + _raise_exceptions_for_connection_errors=False, + **hub_kwargs, + ) + commit_hash = extract_commit_hash(resolved_config_file, commit_hash) + else: + commit_hash = getattr(config, "_commit_hash", None) + + if is_peft_available(): + if adapter_kwargs is None: + adapter_kwargs = {} + if token is not None: + adapter_kwargs["token"] = token + + maybe_adapter_path = find_adapter_config_file( + pretrained_model_name_or_path, _commit_hash=commit_hash, **adapter_kwargs + ) + + if maybe_adapter_path is not None: + with open(maybe_adapter_path, "r", encoding="utf-8") as f: + adapter_config = json.load(f) + + adapter_kwargs["_adapter_model_path"] = pretrained_model_name_or_path + pretrained_model_name_or_path = adapter_config["base_model_name_or_path"] + + if not isinstance(config, PretrainedConfig): + kwargs_orig = copy.deepcopy(kwargs) + # ensure not to pollute the config object with torch_dtype="auto" - since it's + # meaningless in the context of the config object - torch.dtype values are acceptable + if kwargs.get("torch_dtype", None) == "auto": + _ = kwargs.pop("torch_dtype") + # to not overwrite the quantization_config if config has a quantization_config + if kwargs.get("quantization_config", None) is not None: + _ = kwargs.pop("quantization_config") + + config, kwargs = AutoConfig.from_pretrained( + pretrained_model_name_or_path, + return_unused_kwargs=True, + trust_remote_code=trust_remote_code, + code_revision=code_revision, + _commit_hash=commit_hash, + **hub_kwargs, + **kwargs, + ) + + # if torch_dtype=auto was passed here, ensure to pass it on + if kwargs_orig.get("torch_dtype", None) == "auto": + kwargs["torch_dtype"] = "auto" + if kwargs_orig.get("quantization_config", None) is not None: + kwargs["quantization_config"] = kwargs_orig["quantization_config"] + + has_remote_code = hasattr(config, "auto_map") and cls.__name__ in config.auto_map + has_local_code = type(config) in cls._model_mapping.keys() + trust_remote_code = resolve_trust_remote_code( + trust_remote_code, pretrained_model_name_or_path, has_local_code, has_remote_code + ) + + # Set the adapter kwargs + kwargs["adapter_kwargs"] = adapter_kwargs + + if has_remote_code and trust_remote_code: + class_ref = config.auto_map[cls.__name__] + model_class = get_class_from_dynamic_module( + class_ref, pretrained_model_name_or_path, code_revision=code_revision, **hub_kwargs, **kwargs + ) + _ = hub_kwargs.pop("code_revision", None) + if os.path.isdir(pretrained_model_name_or_path): + model_class.register_for_auto_class(cls.__name__) + else: + cls.register(config.__class__, model_class, exist_ok=True) + return model_class.from_pretrained( + pretrained_model_name_or_path, *model_args, config=config, **hub_kwargs, **kwargs + ) + elif type(config) in cls._model_mapping.keys(): + model_class = _get_model_class(config, cls._model_mapping) + return model_class.from_pretrained( + pretrained_model_name_or_path, *model_args, config=config, **hub_kwargs, **kwargs + ) + raise ValueError( + f"Unrecognized configuration class {config.__class__} for this kind of AutoModel: {cls.__name__}.\n" + f"Model type should be one of {', '.join(c.__name__ for c in cls._model_mapping.keys())}." + ) + + @classmethod + def register(cls, config_class, model_class, exist_ok=False): + """ + Register a new model for this class. + + Args: + config_class ([`PretrainedConfig`]): + The configuration corresponding to the model to register. + model_class ([`PreTrainedModel`]): + The model to register. + """ + if hasattr(model_class, "config_class") and str(model_class.config_class) != str(config_class): + raise ValueError( + "The model class you are passing has a `config_class` attribute that is not consistent with the " + f"config class you passed (model has {model_class.config_class} and you passed {config_class}. Fix " + "one of those so they match!" + ) + cls._model_mapping.register(config_class, model_class, exist_ok=exist_ok) + + +class _BaseAutoBackboneClass(_BaseAutoModelClass): + # Base class for auto backbone models. + _model_mapping = None + + @classmethod + def _load_timm_backbone_from_pretrained(cls, pretrained_model_name_or_path, *model_args, **kwargs): + requires_backends(cls, ["vision", "timm"]) + from ...models.timm_backbone import TimmBackboneConfig + + config = kwargs.pop("config", TimmBackboneConfig()) + + if kwargs.get("out_features", None) is not None: + raise ValueError("Cannot specify `out_features` for timm backbones") + + if kwargs.get("output_loading_info", False): + raise ValueError("Cannot specify `output_loading_info=True` when loading from timm") + + num_channels = kwargs.pop("num_channels", config.num_channels) + features_only = kwargs.pop("features_only", config.features_only) + use_pretrained_backbone = kwargs.pop("use_pretrained_backbone", config.use_pretrained_backbone) + out_indices = kwargs.pop("out_indices", config.out_indices) + config = TimmBackboneConfig( + backbone=pretrained_model_name_or_path, + num_channels=num_channels, + features_only=features_only, + use_pretrained_backbone=use_pretrained_backbone, + out_indices=out_indices, + ) + return super().from_config(config, **kwargs) + + @classmethod + def from_pretrained(cls, pretrained_model_name_or_path, *model_args, **kwargs): + use_timm_backbone = kwargs.pop("use_timm_backbone", False) + if use_timm_backbone: + return cls._load_timm_backbone_from_pretrained(pretrained_model_name_or_path, *model_args, **kwargs) + + return super().from_pretrained(pretrained_model_name_or_path, *model_args, **kwargs) + + +def insert_head_doc(docstring, head_doc=""): + if len(head_doc) > 0: + return docstring.replace( + "one of the model classes of the library ", + f"one of the model classes of the library (with a {head_doc} head) ", + ) + return docstring.replace( + "one of the model classes of the library ", "one of the base model classes of the library " + ) + + +def auto_class_update(cls, checkpoint_for_example="google-bert/bert-base-cased", head_doc=""): + # Create a new class with the right name from the base class + model_mapping = cls._model_mapping + name = cls.__name__ + class_docstring = insert_head_doc(CLASS_DOCSTRING, head_doc=head_doc) + cls.__doc__ = class_docstring.replace("BaseAutoModelClass", name) + + # Now we need to copy and re-register `from_config` and `from_pretrained` as class methods otherwise we can't + # have a specific docstrings for them. + from_config = copy_func(_BaseAutoModelClass.from_config) + from_config_docstring = insert_head_doc(FROM_CONFIG_DOCSTRING, head_doc=head_doc) + from_config_docstring = from_config_docstring.replace("BaseAutoModelClass", name) + from_config_docstring = from_config_docstring.replace("checkpoint_placeholder", checkpoint_for_example) + from_config.__doc__ = from_config_docstring + from_config = replace_list_option_in_docstrings(model_mapping._model_mapping, use_model_types=False)(from_config) + cls.from_config = classmethod(from_config) + + if name.startswith("TF"): + from_pretrained_docstring = FROM_PRETRAINED_TF_DOCSTRING + elif name.startswith("Flax"): + from_pretrained_docstring = FROM_PRETRAINED_FLAX_DOCSTRING + else: + from_pretrained_docstring = FROM_PRETRAINED_TORCH_DOCSTRING + from_pretrained = copy_func(_BaseAutoModelClass.from_pretrained) + from_pretrained_docstring = insert_head_doc(from_pretrained_docstring, head_doc=head_doc) + from_pretrained_docstring = from_pretrained_docstring.replace("BaseAutoModelClass", name) + from_pretrained_docstring = from_pretrained_docstring.replace("checkpoint_placeholder", checkpoint_for_example) + shortcut = checkpoint_for_example.split("/")[-1].split("-")[0] + from_pretrained_docstring = from_pretrained_docstring.replace("shortcut_placeholder", shortcut) + from_pretrained.__doc__ = from_pretrained_docstring + from_pretrained = replace_list_option_in_docstrings(model_mapping._model_mapping)(from_pretrained) + cls.from_pretrained = classmethod(from_pretrained) + return cls + + +def get_values(model_mapping): + result = [] + for model in model_mapping.values(): + if isinstance(model, (list, tuple)): + result += list(model) + else: + result.append(model) + + return result + + +def getattribute_from_module(module, attr): + if attr is None: + return None + if isinstance(attr, tuple): + return tuple(getattribute_from_module(module, a) for a in attr) + if hasattr(module, attr): + return getattr(module, attr) + # Some of the mappings have entries model_type -> object of another model type. In that case we try to grab the + # object at the top level. + transformers_module = importlib.import_module("transformers") + + if module != transformers_module: + try: + return getattribute_from_module(transformers_module, attr) + except ValueError: + raise ValueError(f"Could not find {attr} neither in {module} nor in {transformers_module}!") + else: + raise ValueError(f"Could not find {attr} in {transformers_module}!") + + +class _LazyAutoMapping(OrderedDict): + """ + " A mapping config to object (model or tokenizer for instance) that will load keys and values when it is accessed. + + Args: + - config_mapping: The map model type to config class + - model_mapping: The map model type to model (or tokenizer) class + """ + + def __init__(self, config_mapping, model_mapping): + self._config_mapping = config_mapping + self._reverse_config_mapping = {v: k for k, v in config_mapping.items()} + self._model_mapping = model_mapping + self._model_mapping._model_mapping = self + self._extra_content = {} + self._modules = {} + + def __len__(self): + common_keys = set(self._config_mapping.keys()).intersection(self._model_mapping.keys()) + return len(common_keys) + len(self._extra_content) + + def __getitem__(self, key): + if key in self._extra_content: + return self._extra_content[key] + model_type = self._reverse_config_mapping[key.__name__] + if model_type in self._model_mapping: + model_name = self._model_mapping[model_type] + return self._load_attr_from_module(model_type, model_name) + + # Maybe there was several model types associated with this config. + model_types = [k for k, v in self._config_mapping.items() if v == key.__name__] + for mtype in model_types: + if mtype in self._model_mapping: + model_name = self._model_mapping[mtype] + return self._load_attr_from_module(mtype, model_name) + raise KeyError(key) + + def _load_attr_from_module(self, model_type, attr): + module_name = model_type_to_module_name(model_type) + if module_name not in self._modules: + self._modules[module_name] = importlib.import_module(f".{module_name}", "transformers.models") + return getattribute_from_module(self._modules[module_name], attr) + + def keys(self): + mapping_keys = [ + self._load_attr_from_module(key, name) + for key, name in self._config_mapping.items() + if key in self._model_mapping.keys() + ] + return mapping_keys + list(self._extra_content.keys()) + + def get(self, key, default): + try: + return self.__getitem__(key) + except KeyError: + return default + + def __bool__(self): + return bool(self.keys()) + + def values(self): + mapping_values = [ + self._load_attr_from_module(key, name) + for key, name in self._model_mapping.items() + if key in self._config_mapping.keys() + ] + return mapping_values + list(self._extra_content.values()) + + def items(self): + mapping_items = [ + ( + self._load_attr_from_module(key, self._config_mapping[key]), + self._load_attr_from_module(key, self._model_mapping[key]), + ) + for key in self._model_mapping.keys() + if key in self._config_mapping.keys() + ] + return mapping_items + list(self._extra_content.items()) + + def __iter__(self): + return iter(self.keys()) + + def __contains__(self, item): + if item in self._extra_content: + return True + if not hasattr(item, "__name__") or item.__name__ not in self._reverse_config_mapping: + return False + model_type = self._reverse_config_mapping[item.__name__] + return model_type in self._model_mapping + + def register(self, key, value, exist_ok=False): + """ + Register a new model in this mapping. + """ + if hasattr(key, "__name__") and key.__name__ in self._reverse_config_mapping: + model_type = self._reverse_config_mapping[key.__name__] + if model_type in self._model_mapping.keys() and not exist_ok: + raise ValueError(f"'{key}' is already used by a Transformers model.") + + self._extra_content[key] = value