diff --git a/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..014eaf9cc922ad610cbc2b7fbb77d321b7c9eab9 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715682653 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/output.log b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..09b4bdd57eb8c227a7f45c17d664b7339f567e79 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/output.log @@ -0,0 +1,42 @@ + +2024-05-14:10:30:54,403 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:10:30:58,879 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:10:30:58,881 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:10:30:58,881 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step120'} +Traceback (most recent call last): + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 398, in cached_file + resolved_file = hf_hub_download( + File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn + validate_repo_id(arg_value) + File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id + raise HFValidationError( +huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/data/cronscript/ckpts//hf_ckpt//global_step120'. Use `repo_type` argument if needed. +The above exception was the direct cause of the following exception: +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/data/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/data/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/data/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 928, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 631, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 686, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 462, in cached_file + raise EnvironmentError( +OSError: Incorrect path_or_model_id: '/data/cronscript/ckpts//hf_ckpt//global_step120'. Please provide either the path to a local folder or the repo_id of a model on the Hub. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..32b043f84e663e09d03ff5c7c1d58559c33d3f9d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T10:30:54.266112", + "startedAt": "2024-05-14T10:30:53.886264", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3394.3469736842103, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3332.668, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3332.543, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3316.868, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3330.683, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3358.507, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3318.849, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 76.92241668701172 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 5}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..75ce4604900706e3d77431b42c240e517f66ab26 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/logs/debug-internal.log @@ -0,0 +1,181 @@ +2024-05-14 10:30:53,898 INFO StreamThr :8236 [internal.py:wandb_internal():85] W&B internal server running at pid: 8236, started at: 2024-05-14 10:30:53.897695 +2024-05-14 10:30:53,900 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: status +2024-05-14 10:30:53,901 INFO WriterThread:8236 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/run-yu37vekm.wandb +2024-05-14 10:30:53,901 DEBUG SenderThread:8236 [sender.py:send():378] send: header +2024-05-14 10:30:53,911 DEBUG SenderThread:8236 [sender.py:send():378] send: run +2024-05-14 10:30:54,125 INFO SenderThread:8236 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files +2024-05-14 10:30:54,125 INFO SenderThread:8236 [sender.py:_start_run_threads():1123] run started: yu37vekm with start time 1715682653.897371 +2024-05-14 10:30:54,132 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 10:30:54,132 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: check_version +2024-05-14 10:30:54,215 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 10:30:54,216 DEBUG HandlerThread:8236 [system_info.py:__init__():26] System info init +2024-05-14 10:30:54,216 DEBUG HandlerThread:8236 [system_info.py:__init__():41] System info init done +2024-05-14 10:30:54,216 INFO HandlerThread:8236 [system_monitor.py:start():194] Starting system monitor +2024-05-14 10:30:54,217 INFO SystemMonitor:8236 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 10:30:54,217 INFO HandlerThread:8236 [system_monitor.py:probe():214] Collecting system info +2024-05-14 10:30:54,217 INFO SystemMonitor:8236 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 10:30:54,217 INFO SystemMonitor:8236 [interfaces.py:start():188] Started disk monitoring +2024-05-14 10:30:54,218 INFO SystemMonitor:8236 [interfaces.py:start():188] Started memory monitoring +2024-05-14 10:30:54,218 INFO SystemMonitor:8236 [interfaces.py:start():188] Started network monitoring +2024-05-14 10:30:54,266 DEBUG HandlerThread:8236 [system_info.py:probe():150] Probing system +2024-05-14 10:30:54,274 DEBUG HandlerThread:8236 [system_info.py:_probe_git():135] Probing git +2024-05-14 10:30:54,299 ERROR HandlerThread:8236 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 10:30:54,299 DEBUG HandlerThread:8236 [system_info.py:_probe_git():143] Probing git done +2024-05-14 10:30:54,300 DEBUG HandlerThread:8236 [system_info.py:probe():198] Probing system done +2024-05-14 10:30:54,300 DEBUG HandlerThread:8236 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T10:30:54.266112', 'startedAt': '2024-05-14T10:30:53.886264', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3394.3469736842103, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3332.668, 'min': 800.0, 'max': 3400.0}, {'current': 3332.543, 'min': 800.0, 'max': 3400.0}, {'current': 3316.868, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3330.683, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3358.507, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3318.849, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 76.92241668701172}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 10:30:54,300 INFO HandlerThread:8236 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 10:30:54,300 INFO HandlerThread:8236 [system_monitor.py:probe():227] Publishing system info +2024-05-14 10:30:54,301 INFO HandlerThread:8236 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 10:30:54,305 DEBUG SenderThread:8236 [sender.py:send():378] send: files +2024-05-14 10:30:54,305 INFO SenderThread:8236 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 10:30:54,400 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 10:30:54,400 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 10:30:54,400 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: python_packages +2024-05-14 10:30:54,401 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: stop_status +2024-05-14 10:30:54,631 DEBUG SenderThread:8236 [sender.py:send():378] send: telemetry +2024-05-14 10:30:54,812 INFO wandb-upload_0:8236 [upload_job.py:push():130] Uploaded file /tmp/tmp7gniqj8owandb/oxues1px-wandb-metadata.json +2024-05-14 10:30:55,127 INFO Thread-12 :8236 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/output.log +2024-05-14 10:30:55,127 INFO Thread-12 :8236 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/requirements.txt +2024-05-14 10:30:55,127 INFO Thread-12 :8236 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-metadata.json +2024-05-14 10:30:57,127 INFO Thread-12 :8236 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/output.log +2024-05-14 10:30:59,882 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:30:59,949 DEBUG SenderThread:8236 [sender.py:send():378] send: exit +2024-05-14 10:30:59,949 INFO SenderThread:8236 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 10:30:59,949 INFO SenderThread:8236 [sender.py:send_exit():587] handling runtime: 5 +2024-05-14 10:30:59,950 INFO SenderThread:8236 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:30:59,950 INFO SenderThread:8236 [sender.py:send_exit():593] send defer +2024-05-14 10:30:59,951 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:59,951 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 10:30:59,951 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:59,951 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 10:30:59,951 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 1 +2024-05-14 10:30:59,951 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:59,951 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 10:30:59,951 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:59,951 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 10:30:59,951 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 2 +2024-05-14 10:30:59,951 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:59,951 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 10:30:59,951 INFO HandlerThread:8236 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 10:30:59,952 INFO HandlerThread:8236 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 10:30:59,952 DEBUG SystemMonitor:8236 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 10:30:59,952 INFO HandlerThread:8236 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 10:30:59,952 DEBUG SystemMonitor:8236 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 10:30:59,952 INFO HandlerThread:8236 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 10:30:59,952 DEBUG SystemMonitor:8236 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 10:30:59,952 INFO HandlerThread:8236 [interfaces.py:finish():200] Joined network monitor +2024-05-14 10:30:59,954 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:59,954 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 10:30:59,954 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 3 +2024-05-14 10:30:59,954 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:59,954 DEBUG SenderThread:8236 [sender.py:send():378] send: stats +2024-05-14 10:30:59,954 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 10:30:59,955 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:59,955 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 10:30:59,955 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 4 +2024-05-14 10:30:59,955 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:59,955 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 10:30:59,955 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:59,955 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 10:30:59,955 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 5 +2024-05-14 10:30:59,955 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:59,955 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 10:30:59,956 DEBUG SenderThread:8236 [sender.py:send():378] send: summary +2024-05-14 10:30:59,956 INFO SenderThread:8236 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 10:30:59,956 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:59,956 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 10:30:59,956 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 6 +2024-05-14 10:30:59,957 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:30:59,957 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 10:30:59,957 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:30:59,957 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 10:30:59,959 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 10:31:00,025 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 7 +2024-05-14 10:31:00,026 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,026 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 10:31:00,026 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,026 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 10:31:00,129 INFO Thread-12 :8236 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/config.yaml +2024-05-14 10:31:00,129 INFO Thread-12 :8236 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-summary.json +2024-05-14 10:31:00,642 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 8 +2024-05-14 10:31:00,642 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,643 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 10:31:00,643 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,643 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 10:31:00,643 INFO SenderThread:8236 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 10:31:00,643 INFO SenderThread:8236 [job_builder.py:_get_source_type():576] no source found +2024-05-14 10:31:00,643 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 9 +2024-05-14 10:31:00,644 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:00,644 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 10:31:00,644 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:00,644 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 10:31:00,644 INFO SenderThread:8236 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 10:31:00,949 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:01,130 INFO SenderThread:8236 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/output.log +2024-05-14 10:31:01,130 INFO SenderThread:8236 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files +2024-05-14 10:31:01,130 INFO SenderThread:8236 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/config.yaml config.yaml +2024-05-14 10:31:01,130 INFO SenderThread:8236 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/requirements.txt requirements.txt +2024-05-14 10:31:01,130 INFO SenderThread:8236 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/output.log output.log +2024-05-14 10:31:01,130 INFO SenderThread:8236 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-summary.json wandb-summary.json +2024-05-14 10:31:01,130 INFO SenderThread:8236 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-metadata.json wandb-metadata.json +2024-05-14 10:31:01,131 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 10 +2024-05-14 10:31:01,131 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:01,131 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:01,131 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 10:31:01,131 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:01,131 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 10:31:01,131 INFO SenderThread:8236 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:31:01,406 INFO wandb-upload_0:8236 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/requirements.txt +2024-05-14 10:31:01,539 INFO wandb-upload_1:8236 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/config.yaml +2024-05-14 10:31:01,609 INFO wandb-upload_2:8236 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/output.log +2024-05-14 10:31:01,610 INFO wandb-upload_3:8236 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/files/wandb-summary.json +2024-05-14 10:31:01,810 INFO Thread-11 (_thread_body):8236 [sender.py:transition_state():613] send defer: 11 +2024-05-14 10:31:01,811 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:01,811 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 10:31:01,811 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:01,812 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 10:31:01,812 INFO SenderThread:8236 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 10:31:01,812 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 12 +2024-05-14 10:31:01,812 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:01,812 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 10:31:01,812 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:01,812 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 10:31:01,812 INFO SenderThread:8236 [file_stream.py:finish():601] file stream finish called +2024-05-14 10:31:01,949 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:02,042 INFO SenderThread:8236 [file_stream.py:finish():605] file stream finish is done +2024-05-14 10:31:02,042 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 13 +2024-05-14 10:31:02,042 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:02,042 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:02,042 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 10:31:02,042 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:02,042 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 10:31:02,042 INFO SenderThread:8236 [sender.py:transition_state():613] send defer: 14 +2024-05-14 10:31:02,042 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: defer +2024-05-14 10:31:02,043 INFO HandlerThread:8236 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 10:31:02,043 DEBUG SenderThread:8236 [sender.py:send():378] send: final +2024-05-14 10:31:02,043 DEBUG SenderThread:8236 [sender.py:send():378] send: footer +2024-05-14 10:31:02,043 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: defer +2024-05-14 10:31:02,043 INFO SenderThread:8236 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 10:31:02,043 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:02,044 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:02,044 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 10:31:02,044 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 10:31:02,044 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 10:31:02,044 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 10:31:02,044 DEBUG SenderThread:8236 [sender.py:send_request():405] send_request: server_info +2024-05-14 10:31:02,044 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 10:31:02,046 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 10:31:02,099 INFO MainThread:8236 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 10:31:02,099 INFO MainThread:8236 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 10:31:02,099 INFO MainThread:8236 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 10:31:02,099 DEBUG HandlerThread:8236 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 10:31:02,099 INFO HandlerThread:8236 [handler.py:finish():882] shutting down handler +2024-05-14 10:31:03,044 INFO WriterThread:8236 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/run-yu37vekm.wandb +2024-05-14 10:31:03,099 INFO SenderThread:8236 [sender.py:finish():1545] shutting down sender +2024-05-14 10:31:03,099 INFO SenderThread:8236 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 10:31:03,099 INFO SenderThread:8236 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..466ce35a2bdc4d594400a8fc8136364654dc1499 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_setup.py:_flush():76] Configure stats pid to 6969 +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 10:30:53,894 WARNING MainThread:6969 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/logs/debug.log +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/logs/debug-internal.log +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_init.py:init():560] calling init triggers +2024-05-14 10:30:53,894 INFO MainThread:6969 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 10:30:53,895 INFO MainThread:6969 [wandb_init.py:init():610] starting backend +2024-05-14 10:30:53,895 INFO MainThread:6969 [wandb_init.py:init():614] setting up manager +2024-05-14 10:30:53,896 INFO MainThread:6969 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 10:30:53,897 INFO MainThread:6969 [wandb_init.py:init():622] backend started and connected +2024-05-14 10:30:53,899 INFO MainThread:6969 [wandb_init.py:init():711] updated telemetry +2024-05-14 10:30:53,910 INFO MainThread:6969 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 10:30:54,131 INFO MainThread:6969 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 10:30:54,210 INFO MainThread:6969 [wandb_run.py:_on_init():2405] got version response +2024-05-14 10:30:54,211 INFO MainThread:6969 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 10:30:54,400 INFO MainThread:6969 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 10:30:54,400 INFO MainThread:6969 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 10:30:54,400 INFO MainThread:6969 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 10:30:54,401 INFO MainThread:6969 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 10:30:54,402 INFO MainThread:6969 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 10:31:03,100 WARNING MsgRouterThr:6969 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/run-yu37vekm.wandb b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/run-yu37vekm.wandb new file mode 100644 index 0000000000000000000000000000000000000000..a8d1ba6b30587a54d9ca51a29838ef735e2e8914 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_103053-yu37vekm/run-yu37vekm.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_164547-pxpzv850/run-pxpzv850.wandb b/lm-evaluation-harness/wandb/run-20240522_164547-pxpzv850/run-pxpzv850.wandb new file mode 100644 index 0000000000000000000000000000000000000000..d81e8b9f40f190056c46bf6536f5fe0093c61754 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_164547-pxpzv850/run-pxpzv850.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..97f2f58bb4bda0aeb624b1ffc061529696696f77 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/logs/debug-internal.log @@ -0,0 +1,137 @@ +2024-05-22 18:49:28,214 INFO StreamThr :1570 [internal.py:wandb_internal():85] W&B internal server running at pid: 1570, started at: 2024-05-22 18:49:28.210056 +2024-05-22 18:49:28,216 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: status +2024-05-22 18:49:28,219 INFO WriterThread:1570 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/run-kt8p2r8k.wandb +2024-05-22 18:49:28,221 DEBUG SenderThread:1570 [sender.py:send():378] send: header +2024-05-22 18:49:28,223 DEBUG SenderThread:1570 [sender.py:send():378] send: run +2024-05-22 18:49:28,482 INFO SenderThread:1570 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files +2024-05-22 18:49:28,482 INFO SenderThread:1570 [sender.py:_start_run_threads():1123] run started: kt8p2r8k with start time 1716403768.210122 +2024-05-22 18:49:28,483 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 18:49:28,484 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: check_version +2024-05-22 18:49:28,640 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 18:49:28,642 DEBUG HandlerThread:1570 [system_info.py:__init__():26] System info init +2024-05-22 18:49:28,643 DEBUG HandlerThread:1570 [system_info.py:__init__():41] System info init done +2024-05-22 18:49:28,643 INFO HandlerThread:1570 [system_monitor.py:start():194] Starting system monitor +2024-05-22 18:49:28,643 INFO SystemMonitor:1570 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 18:49:28,643 INFO HandlerThread:1570 [system_monitor.py:probe():214] Collecting system info +2024-05-22 18:49:28,650 INFO SystemMonitor:1570 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 18:49:28,650 INFO SystemMonitor:1570 [interfaces.py:start():188] Started disk monitoring +2024-05-22 18:49:28,652 INFO SystemMonitor:1570 [interfaces.py:start():188] Started memory monitoring +2024-05-22 18:49:28,652 INFO SystemMonitor:1570 [interfaces.py:start():188] Started network monitoring +2024-05-22 18:49:28,717 DEBUG HandlerThread:1570 [system_info.py:probe():150] Probing system +2024-05-22 18:49:28,720 DEBUG HandlerThread:1570 [system_info.py:_probe_git():135] Probing git +2024-05-22 18:49:28,730 ERROR HandlerThread:1570 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 18:49:28,730 DEBUG HandlerThread:1570 [system_info.py:_probe_git():143] Probing git done +2024-05-22 18:49:28,730 DEBUG HandlerThread:1570 [system_info.py:probe():198] Probing system done +2024-05-22 18:49:28,730 DEBUG HandlerThread:1570 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T18:49:28.717571', 'startedAt': '2024-05-22T18:49:28.186831', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.1428562499996, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64437866210938}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 18:49:28,731 INFO HandlerThread:1570 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 18:49:28,731 INFO HandlerThread:1570 [system_monitor.py:probe():227] Publishing system info +2024-05-22 18:49:28,733 INFO HandlerThread:1570 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 18:49:28,739 DEBUG SenderThread:1570 [sender.py:send():378] send: files +2024-05-22 18:49:28,739 INFO SenderThread:1570 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 18:49:28,914 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 18:49:28,914 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: python_packages +2024-05-22 18:49:28,916 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 18:49:28,917 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: stop_status +2024-05-22 18:49:29,020 DEBUG SenderThread:1570 [sender.py:send():378] send: telemetry +2024-05-22 18:49:29,397 INFO wandb-upload_0:1570 [upload_job.py:push():130] Uploaded file /tmp/tmpisanswpdwandb/k50sf3dt-wandb-metadata.json +2024-05-22 18:49:29,484 INFO Thread-12 :1570 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/wandb-metadata.json +2024-05-22 18:49:29,484 INFO Thread-12 :1570 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/output.log +2024-05-22 18:49:29,484 INFO Thread-12 :1570 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/requirements.txt +2024-05-22 18:49:31,484 INFO Thread-12 :1570 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/output.log +2024-05-22 18:49:34,023 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:49:39,336 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:49:39,495 INFO Thread-12 :1570 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/output.log +2024-05-22 18:49:39,761 DEBUG SenderThread:1570 [sender.py:send():378] send: exit +2024-05-22 18:49:39,761 INFO SenderThread:1570 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 18:49:39,761 INFO SenderThread:1570 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 18:49:39,762 INFO SenderThread:1570 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:49:39,762 INFO SenderThread:1570 [sender.py:send_exit():593] send defer +2024-05-22 18:49:39,763 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:39,763 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 18:49:39,763 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:39,763 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 18:49:39,763 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 1 +2024-05-22 18:49:39,763 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:39,763 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 18:49:39,763 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:39,763 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 18:49:39,763 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 2 +2024-05-22 18:49:39,763 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:39,763 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 18:49:39,763 INFO HandlerThread:1570 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 18:49:39,763 DEBUG SystemMonitor:1570 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 18:49:39,763 DEBUG SystemMonitor:1570 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 18:49:39,763 DEBUG SystemMonitor:1570 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 18:49:39,764 INFO HandlerThread:1570 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 18:49:39,764 INFO HandlerThread:1570 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 18:49:39,764 INFO HandlerThread:1570 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 18:49:39,765 INFO HandlerThread:1570 [interfaces.py:finish():200] Joined network monitor +2024-05-22 18:49:39,765 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:39,765 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 18:49:39,765 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 3 +2024-05-22 18:49:39,765 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:39,765 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 18:49:39,765 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:39,765 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 18:49:39,765 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 4 +2024-05-22 18:49:39,765 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:39,765 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 18:49:39,765 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:39,765 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 18:49:39,765 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 5 +2024-05-22 18:49:39,765 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:39,765 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 18:49:39,766 DEBUG SenderThread:1570 [sender.py:send():378] send: summary +2024-05-22 18:49:39,766 INFO SenderThread:1570 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:49:39,767 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:39,767 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 18:49:39,767 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 6 +2024-05-22 18:49:39,767 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:39,767 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 18:49:39,767 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:39,767 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 18:49:39,771 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:49:39,865 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 7 +2024-05-22 18:49:39,865 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:39,865 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 18:49:39,865 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:39,865 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 18:49:40,496 INFO Thread-12 :1570 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/config.yaml +2024-05-22 18:49:40,496 INFO Thread-12 :1570 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/wandb-summary.json +2024-05-22 18:49:40,761 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:49:41,045 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 8 +2024-05-22 18:49:41,046 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:49:41,046 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:41,046 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 18:49:41,046 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:41,046 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 18:49:41,046 INFO SenderThread:1570 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 18:49:41,047 INFO SenderThread:1570 [job_builder.py:_get_source_type():576] no source found +2024-05-22 18:49:41,047 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 9 +2024-05-22 18:49:41,047 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:41,047 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 18:49:41,047 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:41,047 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 18:49:41,047 INFO SenderThread:1570 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 18:49:41,498 INFO SenderThread:1570 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/output.log +2024-05-22 18:49:41,498 INFO SenderThread:1570 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files +2024-05-22 18:49:41,498 INFO SenderThread:1570 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/wandb-metadata.json wandb-metadata.json +2024-05-22 18:49:41,498 INFO SenderThread:1570 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/requirements.txt requirements.txt +2024-05-22 18:49:41,498 INFO SenderThread:1570 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/wandb-summary.json wandb-summary.json +2024-05-22 18:49:41,501 INFO SenderThread:1570 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/output.log output.log +2024-05-22 18:49:41,503 INFO SenderThread:1570 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/config.yaml config.yaml +2024-05-22 18:49:41,503 INFO SenderThread:1570 [sender.py:transition_state():613] send defer: 10 +2024-05-22 18:49:41,503 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:49:41,503 INFO HandlerThread:1570 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 18:49:41,503 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: defer +2024-05-22 18:49:41,503 INFO SenderThread:1570 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 18:49:41,503 INFO SenderThread:1570 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:49:41,752 INFO wandb-upload_0:1570 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/requirements.txt +2024-05-22 18:49:41,761 DEBUG HandlerThread:1570 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:49:41,761 DEBUG SenderThread:1570 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:49:42,137 INFO wandb-upload_2:1570 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/files/output.log diff --git a/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c8a8b16fcc8ccc44f985e41260862e71e0de2522 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/logs/debug.log @@ -0,0 +1,28 @@ +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_setup.py:_flush():76] Configure stats pid to 1415 +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 18:49:28,204 WARNING MainThread:1415 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/logs/debug.log +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_184928-kt8p2r8k/logs/debug-internal.log +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_init.py:init():560] calling init triggers +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_init.py:init():610] starting backend +2024-05-22 18:49:28,204 INFO MainThread:1415 [wandb_init.py:init():614] setting up manager +2024-05-22 18:49:28,209 INFO MainThread:1415 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 18:49:28,209 INFO MainThread:1415 [wandb_init.py:init():622] backend started and connected +2024-05-22 18:49:28,213 INFO MainThread:1415 [wandb_init.py:init():711] updated telemetry +2024-05-22 18:49:28,223 INFO MainThread:1415 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 18:49:28,483 INFO MainThread:1415 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 18:49:28,634 INFO MainThread:1415 [wandb_run.py:_on_init():2405] got version response +2024-05-22 18:49:28,634 INFO MainThread:1415 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 18:49:28,915 INFO MainThread:1415 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 18:49:28,915 INFO MainThread:1415 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 18:49:28,915 INFO MainThread:1415 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 18:49:28,915 INFO MainThread:1415 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 18:49:28,918 INFO MainThread:1415 [wandb_init.py:init():838] run started, returning control to user process diff --git a/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..1bb0a137e6a80ff992d188642c421ce5f7ed7b66 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716441730 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..dbc5a5fa56add0840bb49ae692addc6df31ac126 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:05:22:11,517 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:05:22:20,014 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:05:22:20,015 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:05:22:20,016 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step6000'} +2024-05-23:05:22:22,340 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step6000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step6000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..80837debc589b1cca3885889c981a3ad271d7ca7 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T05:22:11.312115", + "startedAt": "2024-05-23T05:22:10.734091", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step6000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2334.21898125, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3333.083, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.237, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 212.24552154541016 + } + }, + "memory": { + "total": 1007.43798828125 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..dd8509160a1ae64b3020470b2648402b3b780c29 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 05:22:10,753 INFO StreamThr :11124 [internal.py:wandb_internal():85] W&B internal server running at pid: 11124, started at: 2024-05-23 05:22:10.751648 +2024-05-23 05:22:10,754 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: status +2024-05-23 05:22:10,759 INFO WriterThread:11124 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/run-qlxjd76q.wandb +2024-05-23 05:22:10,764 DEBUG SenderThread:11124 [sender.py:send():378] send: header +2024-05-23 05:22:10,768 DEBUG SenderThread:11124 [sender.py:send():378] send: run +2024-05-23 05:22:11,073 INFO SenderThread:11124 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files +2024-05-23 05:22:11,074 INFO SenderThread:11124 [sender.py:_start_run_threads():1123] run started: qlxjd76q with start time 1716441730.75226 +2024-05-23 05:22:11,077 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 05:22:11,078 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: check_version +2024-05-23 05:22:11,196 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 05:22:11,199 DEBUG HandlerThread:11124 [system_info.py:__init__():26] System info init +2024-05-23 05:22:11,199 DEBUG HandlerThread:11124 [system_info.py:__init__():41] System info init done +2024-05-23 05:22:11,199 INFO HandlerThread:11124 [system_monitor.py:start():194] Starting system monitor +2024-05-23 05:22:11,199 INFO SystemMonitor:11124 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 05:22:11,199 INFO HandlerThread:11124 [system_monitor.py:probe():214] Collecting system info +2024-05-23 05:22:11,206 INFO SystemMonitor:11124 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 05:22:11,206 INFO SystemMonitor:11124 [interfaces.py:start():188] Started disk monitoring +2024-05-23 05:22:11,206 INFO SystemMonitor:11124 [interfaces.py:start():188] Started memory monitoring +2024-05-23 05:22:11,214 INFO SystemMonitor:11124 [interfaces.py:start():188] Started network monitoring +2024-05-23 05:22:11,312 DEBUG HandlerThread:11124 [system_info.py:probe():150] Probing system +2024-05-23 05:22:11,316 DEBUG HandlerThread:11124 [system_info.py:_probe_git():135] Probing git +2024-05-23 05:22:11,326 ERROR HandlerThread:11124 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 05:22:11,326 DEBUG HandlerThread:11124 [system_info.py:_probe_git():143] Probing git done +2024-05-23 05:22:11,326 DEBUG HandlerThread:11124 [system_info.py:probe():198] Probing system done +2024-05-23 05:22:11,326 DEBUG HandlerThread:11124 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T05:22:11.312115', 'startedAt': '2024-05-23T05:22:10.734091', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step6000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2334.21898125, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3333.083, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.237, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 212.24552154541016}}, 'memory': {'total': 1007.43798828125}} +2024-05-23 05:22:11,326 INFO HandlerThread:11124 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 05:22:11,326 INFO HandlerThread:11124 [system_monitor.py:probe():227] Publishing system info +2024-05-23 05:22:11,329 INFO HandlerThread:11124 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 05:22:11,334 DEBUG SenderThread:11124 [sender.py:send():378] send: files +2024-05-23 05:22:11,334 INFO SenderThread:11124 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 05:22:11,511 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 05:22:11,511 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: python_packages +2024-05-23 05:22:11,512 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 05:22:11,512 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: stop_status +2024-05-23 05:22:11,629 DEBUG SenderThread:11124 [sender.py:send():378] send: telemetry +2024-05-23 05:22:11,888 INFO wandb-upload_0:11124 [upload_job.py:push():130] Uploaded file /tmp/tmpxfu0gkt0wandb/yw4erfpx-wandb-metadata.json +2024-05-23 05:22:12,076 INFO Thread-12 :11124 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log +2024-05-23 05:22:12,076 INFO Thread-12 :11124 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/requirements.txt +2024-05-23 05:22:12,076 INFO Thread-12 :11124 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-metadata.json +2024-05-23 05:22:14,076 INFO Thread-12 :11124 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log +2024-05-23 05:22:16,633 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:22:22,017 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:22:22,083 INFO Thread-12 :11124 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log +2024-05-23 05:22:22,347 DEBUG SenderThread:11124 [sender.py:send():378] send: exit +2024-05-23 05:22:22,348 INFO SenderThread:11124 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 05:22:22,348 INFO SenderThread:11124 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 05:22:22,349 INFO SenderThread:11124 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:22:22,349 INFO SenderThread:11124 [sender.py:send_exit():593] send defer +2024-05-23 05:22:22,349 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:22,349 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 05:22:22,350 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:22,350 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 05:22:22,350 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 1 +2024-05-23 05:22:22,350 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:22,350 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 05:22:22,350 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:22,350 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 05:22:22,350 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 2 +2024-05-23 05:22:22,350 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:22,350 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 05:22:22,350 INFO HandlerThread:11124 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 05:22:22,350 DEBUG SystemMonitor:11124 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 05:22:22,351 DEBUG SystemMonitor:11124 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 05:22:22,351 INFO HandlerThread:11124 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 05:22:22,351 DEBUG SystemMonitor:11124 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 05:22:22,351 INFO HandlerThread:11124 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 05:22:22,353 INFO HandlerThread:11124 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 05:22:22,353 INFO HandlerThread:11124 [interfaces.py:finish():200] Joined network monitor +2024-05-23 05:22:22,353 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:22,353 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 05:22:22,353 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 3 +2024-05-23 05:22:22,353 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:22,353 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 05:22:22,354 DEBUG SenderThread:11124 [sender.py:send():378] send: stats +2024-05-23 05:22:22,355 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:22,355 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 05:22:22,355 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 4 +2024-05-23 05:22:22,355 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:22,355 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 05:22:22,355 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:22,355 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 05:22:22,355 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 5 +2024-05-23 05:22:22,355 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:22,355 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 05:22:22,355 DEBUG SenderThread:11124 [sender.py:send():378] send: summary +2024-05-23 05:22:22,356 INFO SenderThread:11124 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:22:22,356 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:22,356 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 05:22:22,356 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 6 +2024-05-23 05:22:22,356 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:22,356 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 05:22:22,357 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:22,357 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 05:22:22,361 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:22:22,448 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 7 +2024-05-23 05:22:22,448 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:22,448 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 05:22:22,448 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:22,448 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 05:22:23,084 INFO Thread-12 :11124 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/config.yaml +2024-05-23 05:22:23,084 INFO Thread-12 :11124 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-summary.json +2024-05-23 05:22:23,347 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:22:23,657 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 8 +2024-05-23 05:22:23,657 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:22:23,657 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:23,658 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 05:22:23,658 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:23,658 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 05:22:23,658 INFO SenderThread:11124 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 05:22:23,658 INFO SenderThread:11124 [job_builder.py:_get_source_type():576] no source found +2024-05-23 05:22:23,658 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 9 +2024-05-23 05:22:23,659 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:23,659 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 05:22:23,659 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:23,659 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 05:22:23,659 INFO SenderThread:11124 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 05:22:24,085 INFO SenderThread:11124 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log +2024-05-23 05:22:24,085 INFO SenderThread:11124 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files +2024-05-23 05:22:24,086 INFO SenderThread:11124 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/requirements.txt requirements.txt +2024-05-23 05:22:24,086 INFO SenderThread:11124 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-metadata.json wandb-metadata.json +2024-05-23 05:22:24,086 INFO SenderThread:11124 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/config.yaml config.yaml +2024-05-23 05:22:24,088 INFO SenderThread:11124 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log output.log +2024-05-23 05:22:24,090 INFO SenderThread:11124 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-summary.json wandb-summary.json +2024-05-23 05:22:24,090 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 10 +2024-05-23 05:22:24,091 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:24,091 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 05:22:24,091 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:24,091 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 05:22:24,091 INFO SenderThread:11124 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:22:24,320 INFO wandb-upload_0:11124 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/requirements.txt +2024-05-23 05:22:24,348 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:22:24,348 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:22:24,668 INFO wandb-upload_3:11124 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/wandb-summary.json +2024-05-23 05:22:24,679 INFO wandb-upload_2:11124 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/output.log +2024-05-23 05:22:24,704 INFO wandb-upload_1:11124 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/files/config.yaml +2024-05-23 05:22:24,904 INFO Thread-11 (_thread_body):11124 [sender.py:transition_state():613] send defer: 11 +2024-05-23 05:22:24,904 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:24,904 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 05:22:24,904 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:24,904 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 05:22:24,904 INFO SenderThread:11124 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 05:22:24,905 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 12 +2024-05-23 05:22:24,905 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:24,905 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 05:22:24,905 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:24,905 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 05:22:24,905 INFO SenderThread:11124 [file_stream.py:finish():601] file stream finish called +2024-05-23 05:22:24,985 INFO SenderThread:11124 [file_stream.py:finish():605] file stream finish is done +2024-05-23 05:22:24,985 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 13 +2024-05-23 05:22:24,985 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:24,985 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 05:22:24,985 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:24,985 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 05:22:24,985 INFO SenderThread:11124 [sender.py:transition_state():613] send defer: 14 +2024-05-23 05:22:24,985 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:22:24,985 INFO HandlerThread:11124 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 05:22:24,985 DEBUG SenderThread:11124 [sender.py:send():378] send: final +2024-05-23 05:22:24,985 DEBUG SenderThread:11124 [sender.py:send():378] send: footer +2024-05-23 05:22:24,985 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: defer +2024-05-23 05:22:24,985 INFO SenderThread:11124 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 05:22:24,986 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:22:24,986 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:22:24,986 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 05:22:24,986 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 05:22:24,986 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 05:22:24,986 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 05:22:24,987 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:22:24,987 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:22:24,987 DEBUG SenderThread:11124 [sender.py:send_request():405] send_request: server_info +2024-05-23 05:22:25,039 INFO MainThread:11124 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 05:22:25,039 INFO MainThread:11124 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 05:22:25,039 INFO MainThread:11124 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 05:22:25,039 DEBUG HandlerThread:11124 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 05:22:25,039 INFO HandlerThread:11124 [handler.py:finish():882] shutting down handler +2024-05-23 05:22:25,987 INFO WriterThread:11124 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/run-qlxjd76q.wandb +2024-05-23 05:22:26,039 INFO SenderThread:11124 [sender.py:finish():1545] shutting down sender +2024-05-23 05:22:26,039 INFO SenderThread:11124 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:22:26,039 INFO SenderThread:11124 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..1c1d0c12d8874908bac48679e107c09d07857776 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_setup.py:_flush():76] Configure stats pid to 10969 +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 05:22:10,746 WARNING MainThread:10969 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/logs/debug.log +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/logs/debug-internal.log +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_init.py:init():560] calling init triggers +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_init.py:init():610] starting backend +2024-05-23 05:22:10,746 INFO MainThread:10969 [wandb_init.py:init():614] setting up manager +2024-05-23 05:22:10,750 INFO MainThread:10969 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 05:22:10,752 INFO MainThread:10969 [wandb_init.py:init():622] backend started and connected +2024-05-23 05:22:10,759 INFO MainThread:10969 [wandb_init.py:init():711] updated telemetry +2024-05-23 05:22:10,767 INFO MainThread:10969 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 05:22:11,077 INFO MainThread:10969 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 05:22:11,190 INFO MainThread:10969 [wandb_run.py:_on_init():2405] got version response +2024-05-23 05:22:11,190 INFO MainThread:10969 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 05:22:11,512 INFO MainThread:10969 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 05:22:11,512 INFO MainThread:10969 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 05:22:11,512 INFO MainThread:10969 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 05:22:11,512 INFO MainThread:10969 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 05:22:11,515 INFO MainThread:10969 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 05:22:26,040 WARNING MsgRouterThr:10969 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/run-qlxjd76q.wandb b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/run-qlxjd76q.wandb new file mode 100644 index 0000000000000000000000000000000000000000..0fef3fb3752d48a82b58dde97573dc0329ce6052 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_052210-qlxjd76q/run-qlxjd76q.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5e9225e6eba37fa59e9492254b314b3f76d61deb --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716451301 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7f9f419cd123a37b3adf215400ea4b8b65e2317c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:08:01:42,093 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:08:01:50,433 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:08:01:50,434 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:08:01:50,434 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step24000'} +2024-05-23:08:01:52,729 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step24000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step24000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..944989ecb7413c81f2c59a54b899d9e95745614a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T08:01:41.881563", + "startedAt": "2024-05-23T08:01:41.399462", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step24000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2325.7879625, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64005279541016 + } + }, + "memory": { + "total": 1007.4379806518555 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..bee5d62e086269ee6a96533f7e3596a539256a9e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 10}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8c3ed571a3c0ab1f4f81d8b70181c5be2fe78065 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/logs/debug-internal.log @@ -0,0 +1,185 @@ +2024-05-23 08:01:41,420 INFO StreamThr :3068 [internal.py:wandb_internal():85] W&B internal server running at pid: 3068, started at: 2024-05-23 08:01:41.418713 +2024-05-23 08:01:41,425 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: status +2024-05-23 08:01:41,425 INFO WriterThread:3068 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/run-2cqsvyos.wandb +2024-05-23 08:01:41,428 DEBUG SenderThread:3068 [sender.py:send():378] send: header +2024-05-23 08:01:41,431 DEBUG SenderThread:3068 [sender.py:send():378] send: run +2024-05-23 08:01:41,684 INFO SenderThread:3068 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files +2024-05-23 08:01:41,684 INFO SenderThread:3068 [sender.py:_start_run_threads():1123] run started: 2cqsvyos with start time 1716451301.418951 +2024-05-23 08:01:41,688 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 08:01:41,688 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: check_version +2024-05-23 08:01:41,806 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 08:01:41,808 DEBUG HandlerThread:3068 [system_info.py:__init__():26] System info init +2024-05-23 08:01:41,808 DEBUG HandlerThread:3068 [system_info.py:__init__():41] System info init done +2024-05-23 08:01:41,808 INFO HandlerThread:3068 [system_monitor.py:start():194] Starting system monitor +2024-05-23 08:01:41,808 INFO SystemMonitor:3068 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 08:01:41,808 INFO HandlerThread:3068 [system_monitor.py:probe():214] Collecting system info +2024-05-23 08:01:41,815 INFO SystemMonitor:3068 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 08:01:41,816 INFO SystemMonitor:3068 [interfaces.py:start():188] Started disk monitoring +2024-05-23 08:01:41,817 INFO SystemMonitor:3068 [interfaces.py:start():188] Started memory monitoring +2024-05-23 08:01:41,818 INFO SystemMonitor:3068 [interfaces.py:start():188] Started network monitoring +2024-05-23 08:01:41,881 DEBUG HandlerThread:3068 [system_info.py:probe():150] Probing system +2024-05-23 08:01:41,884 DEBUG HandlerThread:3068 [system_info.py:_probe_git():135] Probing git +2024-05-23 08:01:41,894 ERROR HandlerThread:3068 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 08:01:41,894 DEBUG HandlerThread:3068 [system_info.py:_probe_git():143] Probing git done +2024-05-23 08:01:41,894 DEBUG HandlerThread:3068 [system_info.py:probe():198] Probing system done +2024-05-23 08:01:41,894 DEBUG HandlerThread:3068 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T08:01:41.881563', 'startedAt': '2024-05-23T08:01:41.399462', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step24000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2325.7879625, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64005279541016}}, 'memory': {'total': 1007.4379806518555}} +2024-05-23 08:01:41,894 INFO HandlerThread:3068 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 08:01:41,895 INFO HandlerThread:3068 [system_monitor.py:probe():227] Publishing system info +2024-05-23 08:01:41,897 INFO HandlerThread:3068 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 08:01:41,902 DEBUG SenderThread:3068 [sender.py:send():378] send: files +2024-05-23 08:01:41,902 INFO SenderThread:3068 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 08:01:42,079 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 08:01:42,079 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: python_packages +2024-05-23 08:01:42,080 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 08:01:42,089 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: stop_status +2024-05-23 08:01:42,255 DEBUG SenderThread:3068 [sender.py:send():378] send: telemetry +2024-05-23 08:01:42,485 INFO wandb-upload_0:3068 [upload_job.py:push():130] Uploaded file /tmp/tmpzhezkosawandb/o4yjpi0c-wandb-metadata.json +2024-05-23 08:01:42,687 INFO Thread-12 :3068 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-metadata.json +2024-05-23 08:01:42,687 INFO Thread-12 :3068 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/requirements.txt +2024-05-23 08:01:42,687 INFO Thread-12 :3068 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log +2024-05-23 08:01:44,686 INFO Thread-12 :3068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log +2024-05-23 08:01:47,257 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:01:52,435 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:01:52,693 INFO Thread-12 :3068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log +2024-05-23 08:01:52,736 DEBUG SenderThread:3068 [sender.py:send():378] send: exit +2024-05-23 08:01:52,736 INFO SenderThread:3068 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 08:01:52,736 INFO SenderThread:3068 [sender.py:send_exit():587] handling runtime: 10 +2024-05-23 08:01:52,738 INFO SenderThread:3068 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 08:01:52,738 INFO SenderThread:3068 [sender.py:send_exit():593] send defer +2024-05-23 08:01:52,738 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:52,738 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 08:01:52,738 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:52,738 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 08:01:52,738 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 1 +2024-05-23 08:01:52,738 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:52,738 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 08:01:52,739 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:52,739 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 08:01:52,739 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 2 +2024-05-23 08:01:52,739 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:52,739 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 08:01:52,739 INFO HandlerThread:3068 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 08:01:52,739 DEBUG SystemMonitor:3068 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 08:01:52,739 DEBUG SystemMonitor:3068 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 08:01:52,739 DEBUG SystemMonitor:3068 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 08:01:52,742 INFO HandlerThread:3068 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 08:01:52,742 INFO HandlerThread:3068 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 08:01:52,742 INFO HandlerThread:3068 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 08:01:52,742 INFO HandlerThread:3068 [interfaces.py:finish():200] Joined network monitor +2024-05-23 08:01:52,743 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:52,743 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 08:01:52,743 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 3 +2024-05-23 08:01:52,743 DEBUG SenderThread:3068 [sender.py:send():378] send: stats +2024-05-23 08:01:52,744 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:52,744 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 08:01:52,744 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:52,744 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 08:01:52,744 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 4 +2024-05-23 08:01:52,744 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:52,744 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 08:01:52,744 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:52,744 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 08:01:52,744 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 5 +2024-05-23 08:01:52,744 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:52,744 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 08:01:52,745 DEBUG SenderThread:3068 [sender.py:send():378] send: summary +2024-05-23 08:01:52,745 INFO SenderThread:3068 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 08:01:52,746 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:52,746 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 08:01:52,746 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 6 +2024-05-23 08:01:52,746 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:52,746 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 08:01:52,746 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:52,746 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 08:01:52,751 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:01:52,826 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 7 +2024-05-23 08:01:52,826 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:52,826 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 08:01:52,826 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:52,826 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 08:01:53,694 INFO Thread-12 :3068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/config.yaml +2024-05-23 08:01:53,694 INFO Thread-12 :3068 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-summary.json +2024-05-23 08:01:53,736 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:01:54,277 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 8 +2024-05-23 08:01:54,277 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:01:54,277 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:54,277 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 08:01:54,277 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:54,277 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 08:01:54,277 INFO SenderThread:3068 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 08:01:54,278 INFO SenderThread:3068 [job_builder.py:_get_source_type():576] no source found +2024-05-23 08:01:54,278 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 9 +2024-05-23 08:01:54,278 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:54,278 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 08:01:54,278 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:54,278 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 08:01:54,278 INFO SenderThread:3068 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 08:01:54,695 INFO SenderThread:3068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log +2024-05-23 08:01:54,696 INFO SenderThread:3068 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files +2024-05-23 08:01:54,696 INFO SenderThread:3068 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-summary.json wandb-summary.json +2024-05-23 08:01:54,696 INFO SenderThread:3068 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/requirements.txt requirements.txt +2024-05-23 08:01:54,698 INFO SenderThread:3068 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-metadata.json wandb-metadata.json +2024-05-23 08:01:54,700 INFO SenderThread:3068 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/config.yaml config.yaml +2024-05-23 08:01:54,701 INFO SenderThread:3068 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log output.log +2024-05-23 08:01:54,701 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 10 +2024-05-23 08:01:54,701 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:54,701 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 08:01:54,701 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:54,701 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 08:01:54,701 INFO SenderThread:3068 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 08:01:54,736 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:01:54,736 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:01:54,948 INFO wandb-upload_0:3068 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/wandb-summary.json +2024-05-23 08:01:55,376 INFO wandb-upload_2:3068 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/config.yaml +2024-05-23 08:01:55,502 INFO wandb-upload_1:3068 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/requirements.txt +2024-05-23 08:01:55,563 INFO wandb-upload_3:3068 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/files/output.log +2024-05-23 08:01:55,736 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:01:55,737 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:01:55,763 INFO Thread-11 (_thread_body):3068 [sender.py:transition_state():613] send defer: 11 +2024-05-23 08:01:55,764 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:55,764 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 08:01:55,764 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:55,764 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 08:01:55,764 INFO SenderThread:3068 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 08:01:55,764 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 12 +2024-05-23 08:01:55,764 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:55,764 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 08:01:55,764 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:55,764 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 08:01:55,764 INFO SenderThread:3068 [file_stream.py:finish():601] file stream finish called +2024-05-23 08:01:55,827 INFO SenderThread:3068 [file_stream.py:finish():605] file stream finish is done +2024-05-23 08:01:55,827 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 13 +2024-05-23 08:01:55,827 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:55,827 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 08:01:55,827 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:55,827 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 08:01:55,827 INFO SenderThread:3068 [sender.py:transition_state():613] send defer: 14 +2024-05-23 08:01:55,827 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:01:55,827 INFO HandlerThread:3068 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 08:01:55,827 DEBUG SenderThread:3068 [sender.py:send():378] send: final +2024-05-23 08:01:55,827 DEBUG SenderThread:3068 [sender.py:send():378] send: footer +2024-05-23 08:01:55,827 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: defer +2024-05-23 08:01:55,828 INFO SenderThread:3068 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 08:01:55,828 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:01:55,828 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:01:55,828 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 08:01:55,828 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 08:01:55,828 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 08:01:55,828 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 08:01:55,829 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:01:55,829 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:01:55,829 DEBUG SenderThread:3068 [sender.py:send_request():405] send_request: server_info +2024-05-23 08:01:55,884 INFO MainThread:3068 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 08:01:55,884 INFO MainThread:3068 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 08:01:55,884 INFO MainThread:3068 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 08:01:55,885 DEBUG HandlerThread:3068 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 08:01:55,885 INFO HandlerThread:3068 [handler.py:finish():882] shutting down handler +2024-05-23 08:01:56,829 INFO WriterThread:3068 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/run-2cqsvyos.wandb +2024-05-23 08:01:56,884 INFO SenderThread:3068 [sender.py:finish():1545] shutting down sender +2024-05-23 08:01:56,884 INFO SenderThread:3068 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 08:01:56,884 INFO SenderThread:3068 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..208a9b9709efe8bd570e82f461c5b9dc7b5fc7c2 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 08:01:41,412 INFO MainThread:2913 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_setup.py:_flush():76] Configure stats pid to 2913 +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 08:01:41,413 WARNING MainThread:2913 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/logs/debug.log +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/logs/debug-internal.log +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_init.py:init():560] calling init triggers +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_init.py:init():610] starting backend +2024-05-23 08:01:41,413 INFO MainThread:2913 [wandb_init.py:init():614] setting up manager +2024-05-23 08:01:41,417 INFO MainThread:2913 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 08:01:41,418 INFO MainThread:2913 [wandb_init.py:init():622] backend started and connected +2024-05-23 08:01:41,422 INFO MainThread:2913 [wandb_init.py:init():711] updated telemetry +2024-05-23 08:01:41,430 INFO MainThread:2913 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 08:01:41,687 INFO MainThread:2913 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 08:01:41,800 INFO MainThread:2913 [wandb_run.py:_on_init():2405] got version response +2024-05-23 08:01:41,800 INFO MainThread:2913 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 08:01:42,080 INFO MainThread:2913 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 08:01:42,081 INFO MainThread:2913 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 08:01:42,081 INFO MainThread:2913 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 08:01:42,081 INFO MainThread:2913 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 08:01:42,090 INFO MainThread:2913 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 08:01:56,886 WARNING MsgRouterThr:2913 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/run-2cqsvyos.wandb b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/run-2cqsvyos.wandb new file mode 100644 index 0000000000000000000000000000000000000000..4d696c067d8f0187351aea7b256e95eca5c22c63 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_080141-2cqsvyos/run-2cqsvyos.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/config.yaml b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..72dfd38e6eda55f78ad50ce40303238ebc71f174 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/config.yaml @@ -0,0 +1,375 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.1 + framework: huggingface + huggingface_version: 4.36.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1717845026 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 2 + - 13 + - 23 + - 62 + 4: 3.10.12 + 5: 0.17.1 + 6: 4.36.2 + 8: + - 5 + 13: linux-x86_64 +task_configs: + desc: null + value: + arc_easy: + task: arc_easy + group: + - ai2_arc + dataset_path: allenai/ai2_arc + dataset_name: ARC-Easy + training_split: train + validation_split: validation + test_split: test + doc_to_text: 'Question: {{question}} + + Answer:' + doc_to_target: '{{choices.label.index(answerKey)}}' + doc_to_choice: '{{choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{question}} + + Answer:' + metadata: + version: 1.0 + boolq: + task: boolq + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: boolq + training_split: train + validation_split: validation + doc_to_text: '{{passage}} + + Question: {{question}}? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: passage + metadata: + version: 2.0 + copa: + task: copa + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: copa + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n # Drop the period\n connector =\ + \ {\n \"cause\": \"because\",\n \"effect\": \"therefore\",\n\ + \ }[doc[\"question\"]]\n return doc[\"premise\"].strip()[:-1] + f\"\ + \ {connector}\"\n" + doc_to_target: "def doc_to_target(doc):\n correct_choice = doc[\"choice1\"\ + ] if doc[\"label\"] == 0 else doc[\"choice2\"]\n # Connect the sentences\n\ + \ return \" \" + convert_choice(correct_choice)\n" + doc_to_choice: "def doc_to_choice(doc):\n return [\" \" + convert_choice(doc[\"\ + choice1\"]), \" \" + convert_choice(doc[\"choice2\"])]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + indic_arc_challenge_hi: + task: indic_arc_challenge_hi + group: Cognitive-Lab/Indic-ARC-Challenge + dataset_path: Cognitive-Lab/Indic-ARC-Challenge + dataset_name: hi + test_split: test + doc_to_text: 'Question: {{translated_question}} + + Answer:' + doc_to_target: '{{translated_choices.label.index(answerKey)}}' + doc_to_choice: '{{translated_choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{translated_question}} + + Answer:' + metadata: + version: 1.0 + indic_arc_easy_hi: + task: indic_arc_easy_hi + group: Cognitive-Lab/Indic-ARC-Easy + dataset_path: Cognitive-Lab/Indic-ARC-Easy + dataset_name: hi + test_split: test + doc_to_text: 'Question: {{translated_question}} + + Answer:' + doc_to_target: '{{translated_choices.label.index(answerKey)}}' + doc_to_choice: '{{translated_choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{translated_question}} + + Answer:' + metadata: + version: 1.0 + indic_boolq_hi: + task: indic_boolq_hi + group: Cognitive-Lab/Indic-BoolQ + dataset_path: Cognitive-Lab/Indic-BoolQ + dataset_name: hi + validation_split: validation + doc_to_text: 'Passage: {translated_passage} + + Question: {translated_question.strip()} + + Answer:' + doc_to_target: answer + doc_to_choice: + - 'true' + - 'false' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + mrpc: + task: mrpc + group: glue + dataset_path: glue + dataset_name: mrpc + training_split: train + validation_split: validation + doc_to_text: 'Sentence 1: {{sentence1}} + + Sentence 2: {{sentence2}} + + Question: Do both sentences mean the same thing? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + - metric: f1 + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + piqa: + task: piqa + dataset_path: piqa + training_split: train + validation_split: validation + doc_to_text: 'Question: {{goal}} + + Answer:' + doc_to_target: label + doc_to_choice: '{{[sol1, sol2]}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: goal + metadata: + version: 1.0 + sst2: + task: sst2 + group: glue + dataset_path: glue + dataset_name: sst2 + training_split: train + validation_split: validation + doc_to_text: '{{sentence}} + + Question: Is this sentence positive or negative? + + Answer:' + doc_to_target: label + doc_to_choice: + - negative + - positive + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + winogrande: + task: winogrande + dataset_path: winogrande + dataset_name: winogrande_xl + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n answer_to_num = {\"1\": 0, \"2\": 1}\n\ + \ return answer_to_num[doc[\"answer\"]]\n" + doc_to_target: "def doc_to_target(doc):\n idx = doc[\"sentence\"].index(\"\ + _\") + 1\n return doc[\"sentence\"][idx:].strip()\n" + doc_to_choice: "def doc_to_choice(doc):\n idx = doc[\"sentence\"].index(\"\ + _\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"\ + sentence\"][:idx] + opt for opt in options]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: sentence + metadata: + version: 1.0 +cli_configs: + desc: null + value: + model: hf + model_args: pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step100000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer + batch_size: auto + batch_sizes: + - 64 + device: null + use_cache: null + limit: null + bootstrap_iters: 100000 + gen_kwargs: null diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json new file mode 100644 index 0000000000000000000000000000000000000000..a257aa4f8ed0186578caf8c977e3caf21f6829d3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json @@ -0,0 +1 @@ +{"columns": ["Tasks", "Version", "Filter", "num_fewshot", "Metric", "Value", "Stderr"], "data": [["winogrande", 1.0, "none", 0, "acc", "0.48539857932123126", "0.0140"], ["sst2", 1.0, "none", 0, "acc", "0.5160550458715596", "0.0169"], ["piqa", 1.0, "none", 0, "acc", "0.5321001088139282", "0.0116"], ["piqa", 1.0, "none", 0, "acc_norm", "0.49347116430903154", "0.0117"], ["mrpc", 1.0, "none", 0, "acc", "0.3161764705882353", "0.0230"], ["mrpc", 1.0, "none", 0, "f1", "0.0", "0.0000"], ["indic_boolq_hi", 1.0, "none", 0, "acc", "0.6217125382262997", "0.0085"], ["indic_arc_easy_hi", 1.0, "none", 0, "acc", "0.25084175084175087", "0.0089"], ["indic_arc_challenge_hi", 1.0, "none", 0, "acc", "0.20733788395904437", "0.0118"], ["copa", 1.0, "none", 0, "acc", "0.58", "0.0496"], ["boolq", 2.0, "none", 0, "acc", "0.38073394495412843", "0.0085"], ["arc_easy", 1.0, "none", 0, "acc", "0.26346801346801346", "0.0090"], ["arc_easy", 1.0, "none", 0, "acc_norm", "0.2668350168350168", "0.0091"]]} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..778a9a5c4dce6e73bf701dd0389bff8b342b0e24 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log @@ -0,0 +1,805 @@ + +2024-06-08:11:10:27,101 INFO [__main__.py:251] Verbosity set to INFO +2024-06-08:11:10:36,720 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'boolq', 'copa', 'indic_arc_challenge_hi', 'indic_arc_easy_hi', 'indic_boolq_hi', 'mrpc', 'piqa', 'sst2', 'winogrande'] +2024-06-08:11:10:36,721 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-06-08:11:10:36,721 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step100000', 'tokenizer': '/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer'} +2024-06-08:11:10:39,088 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. +Downloading readme: 100%|██████████| 9.00k/9.00k [00:00<00:00, 16.1MB/s] +Downloading data: 100%|██████████| 331k/331k [00:00<00:00, 1.48MB/s] +Downloading data: 100%|██████████| 346k/346k [00:00<00:00, 2.71MB/s] +Downloading data: 100%|██████████| 86.1k/86.1k [00:00<00:00, 706kB/s] +Generating train split: 100%|██████████| 2251/2251 [00:00<00:00, 47691.92 examples/s] +Generating test split: 100%|██████████| 2376/2376 [00:00<00:00, 328325.58 examples/s] +Generating validation split: 100%|██████████| 570/570 [00:00<00:00, 154162.58 examples/s] +2024-06-08:11:11:07,210 WARNING [task.py:763] [Task: boolq] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-08:11:11:07,211 WARNING [task.py:775] [Task: boolq] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for super_glue contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/super_glue +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 30.7k/30.7k [00:00<00:00, 38.8MB/s] +Downloading readme: 100%|██████████| 18.2k/18.2k [00:00<00:00, 26.7MB/s] +Downloading data: 100%|██████████| 4.12M/4.12M [00:00<00:00, 25.5MB/s] +Generating train split: 100%|██████████| 9427/9427 [00:00<00:00, 22364.23 examples/s] +Generating validation split: 100%|██████████| 3270/3270 [00:00<00:00, 22504.25 examples/s] +Generating test split: 100%|██████████| 3245/3245 [00:00<00:00, 22795.47 examples/s] +2024-06-08:11:11:10,503 WARNING [task.py:763] [Task: copa] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-08:11:11:10,504 WARNING [task.py:775] [Task: copa] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +Downloading data: 100%|██████████| 44.0k/44.0k [00:00<00:00, 43.7MB/s] +Generating train split: 100%|██████████| 400/400 [00:00<00:00, 16554.72 examples/s] +Generating validation split: 100%|██████████| 100/100 [00:00<00:00, 13212.90 examples/s] +Generating test split: 100%|██████████| 500/500 [00:00<00:00, 16539.31 examples/s] +Downloading readme: 100%|██████████| 2.09k/2.09k [00:00<00:00, 5.02MB/s] +Downloading data: 100%|██████████| 1.73M/1.73M [00:00<00:00, 5.42MB/s] +Downloading data: 100%|██████████| 1.84M/1.84M [00:00<00:00, 5.89MB/s] +Downloading data: 100%|██████████| 473k/473k [00:00<00:00, 1.94MB/s] +Generating train split: 1119 examples [00:00, 29847.10 examples/s] +Generating test split: 1172 examples [00:00, 24003.26 examples/s] +Generating validation split: 299 examples [00:00, 9589.73 examples/s] +2024-06-08:11:11:15,235 WARNING [task.py:322] [Task: indic_arc_challenge_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-08:11:11:15,236 WARNING [task.py:322] [Task: indic_arc_challenge_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +Downloading readme: 100%|██████████| 2.19k/2.19k [00:00<00:00, 12.7MB/s] +Downloading data: 100%|██████████| 3.18M/3.18M [00:00<00:00, 15.8MB/s] +Downloading data: 100%|██████████| 3.37M/3.37M [00:00<00:00, 7.79MB/s] +Downloading data: 100%|██████████| 808k/808k [00:00<00:00, 8.55MB/s] +Generating train split: 100%|██████████| 2251/2251 [00:00<00:00, 37212.97 examples/s] +Generating test split: 100%|██████████| 2376/2376 [00:00<00:00, 32758.20 examples/s] +Generating validation split: 100%|██████████| 570/570 [00:00<00:00, 17717.03 examples/s] +2024-06-08:11:11:18,419 WARNING [task.py:322] [Task: indic_arc_easy_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-08:11:11:18,419 WARNING [task.py:322] [Task: indic_arc_easy_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +Downloading readme: 100%|██████████| 1.53k/1.53k [00:00<00:00, 9.19MB/s] +Downloading data: 100%|██████████| 22.1M/22.1M [00:01<00:00, 18.4MB/s] +Downloading data: 100%|██████████| 7.55M/7.55M [00:00<00:00, 15.9MB/s] +Generating train split: 100%|██████████| 9427/9427 [00:00<00:00, 36151.38 examples/s] +Generating validation split: 100%|██████████| 3270/3270 [00:00<00:00, 35974.85 examples/s] +2024-06-08:11:11:22,746 WARNING [task.py:763] [Task: mrpc] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-08:11:11:22,746 WARNING [task.py:775] [Task: mrpc] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +2024-06-08:11:11:22,746 WARNING [task.py:763] [Task: mrpc] metric f1 is defined, but aggregation is not. using default aggregation=f1 +2024-06-08:11:11:22,747 WARNING [task.py:775] [Task: mrpc] metric f1 is defined, but higher_is_better is not. using default higher_is_better=True +Downloading readme: 100%|██████████| 35.3k/35.3k [00:00<00:00, 63.6MB/s] +Downloading data: 100%|██████████| 649k/649k [00:00<00:00, 3.52MB/s] +Downloading data: 100%|██████████| 75.7k/75.7k [00:00<00:00, 527kB/s] +Downloading data: 100%|██████████| 308k/308k [00:00<00:00, 2.12MB/s] +Generating train split: 100%|██████████| 3668/3668 [00:00<00:00, 415140.91 examples/s] +Generating validation split: 100%|██████████| 408/408 [00:00<00:00, 181260.04 examples/s] +Generating test split: 100%|██████████| 1725/1725 [00:00<00:00, 387924.21 examples/s] +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for piqa contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/piqa +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 5.36k/5.36k [00:00<00:00, 22.5MB/s] +Downloading readme: 100%|██████████| 8.41k/8.41k [00:00<00:00, 31.5MB/s] +Downloading data: 100%|██████████| 1.82M/1.82M [00:00<00:00, 4.19MB/s] +Downloading data: 100%|██████████| 815k/815k [00:00<00:00, 53.8MB/s] +Generating train split: 100%|██████████| 16113/16113 [00:00<00:00, 23818.46 examples/s] +Generating test split: 100%|██████████| 3084/3084 [00:00<00:00, 24787.45 examples/s] +Generating validation split: 100%|██████████| 1838/1838 [00:00<00:00, 23782.31 examples/s] +2024-06-08:11:11:32,824 WARNING [task.py:763] [Task: sst2] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-08:11:11:32,825 WARNING [task.py:775] [Task: sst2] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +Downloading data: 100%|██████████| 3.11M/3.11M [00:00<00:00, 20.6MB/s] +Downloading data: 100%|██████████| 72.8k/72.8k [00:00<00:00, 508kB/s] +Downloading data: 100%|██████████| 148k/148k [00:00<00:00, 1.02MB/s] +Generating train split: 100%|██████████| 67349/67349 [00:00<00:00, 1414143.23 examples/s] +Generating validation split: 100%|██████████| 872/872 [00:00<00:00, 372827.02 examples/s] +Generating test split: 100%|██████████| 1821/1821 [00:00<00:00, 598998.32 examples/s] +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for winogrande contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/winogrande +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 5.65k/5.65k [00:00<00:00, 24.7MB/s] +Downloading readme: 100%|██████████| 9.97k/9.97k [00:00<00:00, 20.7MB/s] +Downloading data: 100%|██████████| 3.40M/3.40M [00:00<00:00, 6.98MB/s] +Generating train split: 100%|██████████| 40398/40398 [00:01<00:00, 23624.41 examples/s] +Generating test split: 100%|██████████| 1767/1767 [00:00<00:00, 24427.28 examples/s] +Generating validation split: 100%|██████████| 1267/1267 [00:00<00:00, 23893.31 examples/s] +2024-06-08:11:11:45,062 INFO [task.py:395] Building contexts for winogrande on rank 0... +100%|██████████| 1267/1267 [00:00<00:00, 68770.65it/s] +2024-06-08:11:11:45,158 INFO [task.py:395] Building contexts for sst2 on rank 0... +100%|██████████| 872/872 [00:00<00:00, 2509.76it/s] +2024-06-08:11:11:45,533 INFO [task.py:395] Building contexts for piqa on rank 0... +100%|██████████| 408/408 [00:00<00:00, 1840.66it/s]s] +2024-06-08:11:11:47,339 INFO [task.py:395] Building contexts for mrpc on rank 0... +100%|██████████| 408/408 [00:00<00:00, 1840.66it/s]s] +2024-06-08:11:11:47,581 INFO [task.py:395] Building contexts for indic_boolq_hi on rank 0... +100%|██████████| 3270/3270 [00:00<00:00, 3677.79it/s] +2024-06-08:11:11:48,639 INFO [task.py:395] Building contexts for indic_arc_easy_hi on rank 0... +100%|██████████| 2376/2376 [00:02<00:00, 1029.09it/s] +2024-06-08:11:11:51,186 INFO [task.py:395] Building contexts for indic_arc_challenge_hi on rank 0... +100%|██████████| 1172/1172 [00:01<00:00, 1125.50it/s] +2024-06-08:11:11:52,347 INFO [task.py:395] Building contexts for copa on rank 0... +100%|██████████| 100/100 [00:00<00:00, 62211.57it/s] +2024-06-08:11:11:52,356 INFO [task.py:395] Building contexts for boolq on rank 0... +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +2024-06-08:11:11:54,145 INFO [task.py:395] Building contexts for arc_easy on rank 0... + 24%|██▍ | 581/2376 [00:00<00:01, 1143.34it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +Passed argument batch_size = auto:1. Detecting largest batch size +Determined largest batch size: 64 +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +bootstrapping for stddev: f1_score +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +100%|██████████| 3270/3270 [00:01<00:00, 1979.26it/s] +hf (pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step100000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer), gen_kwargs: (None), limit: None, num_fewshot: None, batch_size: auto (64) +| Tasks |Version|Filter|n-shot| Metric |Value | |Stderr| +|----------------------|------:|------|-----:|--------|-----:|---|-----:| +|winogrande | 1|none | 0|acc |0.4854|± |0.0140| +|sst2 | 1|none | 0|acc |0.5161|± |0.0169| +|piqa | 1|none | 0|acc |0.5321|± |0.0116| +| | |none | 0|acc_norm|0.4935|± |0.0117| +|mrpc | 1|none | 0|acc |0.3162|± |0.0230| +| | |none | 0|f1 |0.0000|± |0.0000| +|indic_boolq_hi | 1|none | 0|acc |0.6217|± |0.0085| +|indic_arc_easy_hi | 1|none | 0|acc |0.2508|± |0.0089| +|indic_arc_challenge_hi| 1|none | 0|acc |0.2073|± |0.0118| +|copa | 1|none | 0|acc |0.5800|± |0.0496| +|boolq | 2|none | 0|acc |0.3807|± |0.0085| +|arc_easy | 1|none | 0|acc |0.2635|± |0.0090| +| | |none | 0|acc_norm|0.2668|± |0.0091| \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a150a62c9ff021a2c5277fcfe678c8af7aa67278 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/requirements.txt @@ -0,0 +1,154 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.31.0 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.2 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.3 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.32.3 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.5.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.15.2 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.36.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.1 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..f0a419ef98230796cc310617619203cc69bba568 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-06-08T11:10:26.878917", + "startedAt": "2024-06-08T11:10:26.443305", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step100000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer", + "--tasks", + "winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq,indic_arc_easy_hi,indic_arc_challenge_hi,indic_boolq_hi", + "--batch_size", + "auto", + "--wandb_args", + "project=engl-hi-eval,name=globalstep_100000,group=exp2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2334.37498125, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 214.5283966064453 + } + }, + "memory": { + "total": 1007.4379501342773 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..21895b48599cf4da9f8a68c6f092a5ce6ba0749b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-summary.json @@ -0,0 +1 @@ +{"winogrande/alias": "winogrande", "sst2/alias": "sst2", "piqa/alias": "piqa", "mrpc/alias": "mrpc", "indic_boolq_hi/alias": "indic_boolq_hi", "indic_arc_easy_hi/alias": "indic_arc_easy_hi", "indic_arc_challenge_hi/alias": "indic_arc_challenge_hi", "copa/alias": "copa", "boolq/alias": "boolq", "arc_easy/alias": "arc_easy", "winogrande/acc": 0.48539857932123126, "winogrande/acc_stderr": 0.014046492383275837, "sst2/acc": 0.5160550458715596, "sst2/acc_stderr": 0.01693311741999099, "piqa/acc": 0.5321001088139282, "piqa/acc_stderr": 0.01164175801482013, "piqa/acc_norm": 0.49347116430903154, "piqa/acc_norm_stderr": 0.01166482959521097, "mrpc/acc": 0.3161764705882353, "mrpc/acc_stderr": 0.023048336668420193, "mrpc/f1": 0.0, "mrpc/f1_stderr": 0.0, "indic_boolq_hi/acc": 0.6217125382262997, "indic_boolq_hi/acc_stderr": 0.008482001133930994, "indic_arc_easy_hi/acc": 0.25084175084175087, "indic_arc_easy_hi/acc_stderr": 0.008895183010487388, "indic_arc_challenge_hi/acc": 0.20733788395904437, "indic_arc_challenge_hi/acc_stderr": 0.011846905782971373, "copa/acc": 0.58, "copa/acc_stderr": 0.049604496374885836, "boolq/acc": 0.38073394495412843, "boolq/acc_stderr": 0.008492625561656218, "arc_easy/acc": 0.26346801346801346, "arc_easy/acc_stderr": 0.009039157374497722, "arc_easy/acc_norm": 0.2668350168350168, "arc_easy/acc_norm_stderr": 0.00907591585926726, "_timestamp": 1717849398.984251, "_runtime": 4372.5210790634155, "_step": 1, "evaluation/eval_results": {"_type": "table-file", "sha256": "6529e3311149275b8699b4a67df73b70fcae908dd4f7d4af4f1c71c1b876ceba", "size": 984, "artifact_path": "wandb-client-artifact://zo4fjo3s3gxn5pzewdndrodktw0wj3vlmo930nos5bjauxgdso83b1v093mu7m3sjuc5l6ifeylay3tou90n1h393ka4vaikp9fzx5ch4qy9hs46r1kolxfyzeg1spz2/evaluation/eval_results.table.json", "_latest_artifact_path": "wandb-client-artifact://c0hwnhshjspl5o1obcimj38jhq5xicsi4rv04t2hvsil0gtkz1fygm2k2h673hytf5f073wg35atyzrch6c7p9zvw230orttwtjxy8zum5c2znzs56jehwecqu6nxcuj:latest/evaluation/eval_results.table.json", "path": "media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json", "ncols": 7, "nrows": 13}, "_wandb": {"runtime": 4374}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..35d0a4d7b66614943c4270749c9ae561ddf7478e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/logs/debug-internal.log @@ -0,0 +1,6803 @@ +2024-06-08 11:10:26,466 INFO StreamThr :989 [internal.py:wandb_internal():85] W&B internal server running at pid: 989, started at: 2024-06-08 11:10:26.462115 +2024-06-08 11:10:26,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status +2024-06-08 11:10:26,469 INFO WriterThread:989 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/run-9apxn9eo.wandb +2024-06-08 11:10:26,472 DEBUG SenderThread:989 [sender.py:send():379] send: header +2024-06-08 11:10:26,476 DEBUG SenderThread:989 [sender.py:send():379] send: run +2024-06-08 11:10:26,719 INFO SenderThread:989 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files +2024-06-08 11:10:26,719 INFO SenderThread:989 [sender.py:_start_run_threads():1188] run started: 9apxn9eo with start time 1717845026.463172 +2024-06-08 11:10:26,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: check_version +2024-06-08 11:10:26,725 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: check_version +2024-06-08 11:10:26,802 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: run_start +2024-06-08 11:10:26,805 DEBUG HandlerThread:989 [system_info.py:__init__():26] System info init +2024-06-08 11:10:26,805 DEBUG HandlerThread:989 [system_info.py:__init__():41] System info init done +2024-06-08 11:10:26,806 INFO HandlerThread:989 [system_monitor.py:start():194] Starting system monitor +2024-06-08 11:10:26,806 INFO SystemMonitor:989 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-06-08 11:10:26,806 INFO HandlerThread:989 [system_monitor.py:probe():214] Collecting system info +2024-06-08 11:10:26,813 INFO SystemMonitor:989 [interfaces.py:start():188] Started cpu monitoring +2024-06-08 11:10:26,813 INFO SystemMonitor:989 [interfaces.py:start():188] Started disk monitoring +2024-06-08 11:10:26,814 INFO SystemMonitor:989 [interfaces.py:start():188] Started memory monitoring +2024-06-08 11:10:26,824 INFO SystemMonitor:989 [interfaces.py:start():188] Started network monitoring +2024-06-08 11:10:26,878 DEBUG HandlerThread:989 [system_info.py:probe():152] Probing system +2024-06-08 11:10:26,881 DEBUG HandlerThread:989 [system_info.py:_probe_git():137] Probing git +2024-06-08 11:10:26,891 ERROR HandlerThread:989 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-06-08 11:10:26,891 DEBUG HandlerThread:989 [system_info.py:_probe_git():145] Probing git done +2024-06-08 11:10:26,891 DEBUG HandlerThread:989 [system_info.py:probe():200] Probing system done +2024-06-08 11:10:26,891 DEBUG HandlerThread:989 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-06-08T11:10:26.878917', 'startedAt': '2024-06-08T11:10:26.443305', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step100000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer', '--tasks', 'winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq,indic_arc_easy_hi,indic_arc_challenge_hi,indic_boolq_hi', '--batch_size', 'auto', '--wandb_args', 'project=engl-hi-eval,name=globalstep_100000,group=exp2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2334.37498125, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.002, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 214.5283966064453}}, 'memory': {'total': 1007.4379501342773}} +2024-06-08 11:10:26,892 INFO HandlerThread:989 [system_monitor.py:probe():224] Finished collecting system info +2024-06-08 11:10:26,892 INFO HandlerThread:989 [system_monitor.py:probe():227] Publishing system info +2024-06-08 11:10:26,896 INFO HandlerThread:989 [system_monitor.py:probe():229] Finished publishing system info +2024-06-08 11:10:26,902 DEBUG SenderThread:989 [sender.py:send():379] send: files +2024-06-08 11:10:26,903 INFO SenderThread:989 [sender.py:_save_file():1454] saving file wandb-metadata.json with policy now +2024-06-08 11:10:27,094 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: python_packages +2024-06-08 11:10:27,094 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: python_packages +2024-06-08 11:10:27,095 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:10:27,097 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:10:27,102 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:27,175 DEBUG SenderThread:989 [sender.py:send():379] send: telemetry +2024-06-08 11:10:27,422 INFO wandb-upload_0:989 [upload_job.py:push():130] Uploaded file /tmp/tmpjx21gu8gwandb/wicntau7-wandb-metadata.json +2024-06-08 11:10:27,721 INFO Thread-12 :989 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-metadata.json +2024-06-08 11:10:27,722 INFO Thread-12 :989 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:10:27,722 INFO Thread-12 :989 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/requirements.txt +2024-06-08 11:10:28,100 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:29,101 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:29,721 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:10:30,101 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:31,102 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:32,095 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:10:32,104 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:33,105 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:34,108 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:35,108 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:36,110 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:37,729 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:10:37,762 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:10:38,892 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:39,732 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:10:39,892 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:40,892 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:41,892 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:42,095 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:10:42,095 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:10:42,892 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:43,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:10:43,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:44,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:45,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:46,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:47,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:48,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:10:48,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:49,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:50,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:51,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:52,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:53,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:10:53,893 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:54,894 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:55,894 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:56,894 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:57,095 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:10:57,095 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:10:57,894 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:58,894 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:10:59,897 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:10:59,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:00,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:01,077 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/config.yaml +2024-06-08 11:11:01,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:02,156 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:02,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:03,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:04,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:05,319 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:05,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:05,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:06,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:07,322 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:07,898 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:08,908 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:09,667 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:09,902 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:10,905 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:11,504 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:11,780 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:11,905 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:12,095 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:11:12,095 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:11:12,908 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:13,786 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:13,908 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:14,912 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:15,788 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:15,912 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:16,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:17,030 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:17,791 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:17,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:18,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:19,793 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:19,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:20,922 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:21,796 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:21,922 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:22,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:22,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:23,802 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:23,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:24,929 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:25,805 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:25,929 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:26,824 DEBUG SystemMonitor:989 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-06-08 11:11:26,827 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:11:26,933 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:27,095 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:11:27,095 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:11:27,242 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:27,807 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:27,933 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:28,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:29,809 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:29,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:30,940 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:31,945 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:32,002 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:32,824 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:32,945 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:33,945 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:34,004 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:34,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:35,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:36,007 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:36,949 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:37,332 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:37,909 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:37,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:38,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:39,500 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:39,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:40,956 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:41,503 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:41,957 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:42,096 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:11:42,096 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:11:42,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:42,993 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:43,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:44,282 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:44,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:45,468 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:45,969 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:46,974 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:47,954 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:47,974 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:48,051 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:49,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:49,459 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:49,984 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:50,990 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:51,995 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:52,231 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:52,998 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:53,098 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:53,814 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:54,003 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:55,008 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:55,988 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:56,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:56,829 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:11:57,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:57,096 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:11:57,096 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:11:58,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:11:58,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:11:58,599 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:11:59,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:00,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:01,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:02,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:03,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:03,913 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:04,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:05,014 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:06,015 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:07,015 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:08,015 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:09,015 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:09,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:10,015 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:11,015 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:11,381 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:12:12,129 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:12:12,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:12:12,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:13,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:13,461 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:12:14,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:14,523 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:12:14,913 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:15,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:16,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:17,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:18,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:19,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:20,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:20,913 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:21,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:22,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:23,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:24,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:25,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:25,913 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:26,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:26,831 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:12:27,129 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:12:27,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:12:27,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:28,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:29,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:30,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:31,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:31,913 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:32,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:33,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:34,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:35,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:36,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:37,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:37,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:38,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:39,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:40,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:41,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:42,129 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:12:42,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:12:42,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:43,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:43,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:44,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:45,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:46,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:47,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:48,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:48,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:49,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:50,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:51,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:52,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:53,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:54,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:54,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:12:55,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:56,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:56,834 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:12:57,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:12:57,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:12:57,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:58,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:12:59,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:00,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:00,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:01,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:02,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:03,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:04,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:05,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:06,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:06,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:07,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:08,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:09,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:10,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:11,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:11,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:12,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:13:12,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:13:12,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:13,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:14,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:15,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:16,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:17,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:17,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:18,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:19,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:20,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:21,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:22,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:23,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:23,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:24,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:25,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:26,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:26,836 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:13:27,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:13:27,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:13:27,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:28,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:28,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:29,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:30,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:31,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:32,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:33,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:33,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:34,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:35,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:36,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:37,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:38,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:39,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:39,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:40,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:41,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:42,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:13:42,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:13:42,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:43,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:44,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:44,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:45,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:46,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:47,139 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:47,815 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:13:48,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:49,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:50,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:50,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:51,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:52,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:53,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:54,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:55,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:56,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:56,838 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:13:56,839 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:13:57,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:13:57,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:13:57,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:58,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:13:59,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:00,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:01,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:01,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:02,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:03,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:03,728 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:14:04,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:05,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:06,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:07,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:07,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:08,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:09,141 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:10,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:11,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:12,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:14:12,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:14:12,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:13,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:13,266 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:13,746 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:14:14,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:15,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:16,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:17,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:18,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:18,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:19,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:20,142 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:21,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:22,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:23,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:23,798 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:14:23,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:24,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:25,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:26,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:26,841 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:14:27,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:14:27,130 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:14:27,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:28,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:29,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:29,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:30,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:31,143 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:32,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:33,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:34,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:34,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:35,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:36,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:37,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:37,842 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:14:38,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:39,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:40,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:40,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:41,144 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:42,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:14:42,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:14:42,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:43,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:44,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:45,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:46,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:46,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:47,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:48,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:49,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:50,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:51,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:52,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:52,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:53,145 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:54,131 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:14:54,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:55,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:56,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:56,842 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:14:57,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:14:57,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:14:57,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:58,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:14:58,291 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:14:59,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:00,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:01,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:02,138 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:15:02,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:03,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:03,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:04,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:05,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:06,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:07,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:08,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:08,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:09,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:10,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:11,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:12,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:15:12,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:15:12,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:13,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:14,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:14,772 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:15:14,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:15,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:16,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:17,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:18,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:19,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:19,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:20,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:21,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:22,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:23,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:24,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:24,405 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:15:25,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:25,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:26,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:26,844 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:15:27,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:15:27,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:15:27,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:28,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:29,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:30,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:30,916 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:31,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:32,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:33,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:34,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:34,834 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:15:35,151 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:35,917 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:36,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:37,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:38,151 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:39,151 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:40,151 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:41,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:41,917 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:42,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:15:42,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:15:42,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:43,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:44,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:45,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:46,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:47,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:47,755 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:47,958 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:15:48,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:49,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:50,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:51,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:52,152 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:52,917 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:53,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:54,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:55,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:56,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:56,846 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:15:57,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:15:57,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:15:57,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:58,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:15:58,289 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:15:59,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:00,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:01,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:02,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:03,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:03,917 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:03,923 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:16:04,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:05,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:06,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:07,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:08,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:09,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:09,917 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:10,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:10,419 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:16:11,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:12,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:16:12,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:16:12,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:13,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:14,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:14,917 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:15,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:16,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:17,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:18,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:19,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:19,917 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:20,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:21,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:22,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:23,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:24,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:24,603 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:16:25,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:25,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:26,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:26,848 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:16:27,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:16:27,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:16:27,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:28,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:29,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:30,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:30,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:31,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:32,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:32,271 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:16:33,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:34,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:34,642 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:16:35,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:36,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:36,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:37,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:38,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:39,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:40,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:41,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:41,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:42,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:16:42,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:16:42,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:43,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:44,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:44,209 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:16:45,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:46,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:47,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:47,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:48,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:49,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:50,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:51,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:52,151 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:16:52,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:53,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:53,369 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:54,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:54,455 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:16:55,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:56,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:56,851 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:16:57,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:16:57,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:16:57,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:58,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:16:58,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:16:59,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:00,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:01,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:02,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:03,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:03,794 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:17:04,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:04,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:05,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:06,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:07,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:08,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:09,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:09,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:10,160 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:11,160 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:12,056 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:17:12,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:17:12,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:17:12,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:13,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:14,122 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:17:14,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:15,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:15,918 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:16,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:17,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:18,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:19,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:20,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:21,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:21,919 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:22,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:23,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:24,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:24,675 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:17:25,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:26,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:26,853 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:17:27,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:17:27,131 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:17:27,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:27,239 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:28,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:29,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:30,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:31,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:32,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:32,994 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:33,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:33,831 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:17:34,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:35,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:36,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:37,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:38,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:38,919 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:39,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:40,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:41,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:42,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:17:42,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:17:42,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:43,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:43,702 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:17:44,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:44,919 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:45,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:46,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:47,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:48,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:49,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:49,919 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:50,169 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:50,575 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:17:51,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:52,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:53,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:54,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:55,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:55,919 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:17:56,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:56,855 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:17:57,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:17:57,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:17:57,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:58,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:17:59,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:00,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:01,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:01,919 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:02,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:03,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:04,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:04,579 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:18:05,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:05,747 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:18:06,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:07,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:07,919 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:08,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:09,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:10,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:11,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:12,131 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:18:12,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:18:12,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:13,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:13,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:14,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:15,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:16,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:16,920 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:18:17,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:18,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:18,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:19,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:20,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:20,254 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:18:21,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:22,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:23,172 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:23,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:24,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:25,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:26,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:26,857 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:18:27,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:18:27,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:18:27,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:28,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:28,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:29,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:30,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:31,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:32,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:33,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:33,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:33,971 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:18:34,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:35,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:36,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:37,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:38,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:38,635 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:18:38,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:39,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:40,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:41,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:42,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:18:42,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:18:42,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:43,174 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:43,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:44,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:45,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:46,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:47,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:48,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:49,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:49,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:50,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:51,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:52,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:53,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:54,004 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:18:54,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:54,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:18:55,175 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:56,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:56,859 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:18:57,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:18:57,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:18:57,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:58,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:59,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:18:59,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:00,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:01,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:02,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:03,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:04,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:04,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:05,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:06,176 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:06,653 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:19:07,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:08,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:09,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:10,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:10,214 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:19:10,920 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:11,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:12,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:19:12,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:19:12,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:13,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:14,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:15,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:15,921 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:16,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:17,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:18,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:19,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:20,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:21,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:21,921 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:22,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:23,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:24,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:24,519 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:19:25,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:26,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:26,791 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:19:26,862 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:19:27,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:19:27,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:19:27,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:27,218 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:28,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:29,178 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:30,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:31,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:32,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:32,921 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:33,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:34,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:35,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:36,040 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:19:36,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:37,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:38,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:38,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:39,179 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:40,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:40,478 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:19:41,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:42,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:19:42,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:19:42,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:43,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:43,239 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:44,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:45,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:46,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:47,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:48,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:48,921 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:49,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:50,180 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:51,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:52,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:53,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:54,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:54,399 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:19:54,921 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:19:55,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:56,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:56,864 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:19:57,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:19:57,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:19:57,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:58,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:59,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:19:59,921 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:00,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:01,181 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:02,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:03,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:04,019 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:20:04,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:05,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:05,921 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:06,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:07,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:08,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:08,795 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:20:09,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:10,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:10,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:11,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:12,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:20:12,132 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:20:12,182 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:13,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:14,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:15,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:16,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:16,922 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:17,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:18,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:19,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:20,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:21,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:21,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:22,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:23,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:24,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:24,595 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:20:25,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:26,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:26,866 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:20:27,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:20:27,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:20:27,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:27,256 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:28,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:29,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:30,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:31,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:32,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:32,922 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:33,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:33,918 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:20:34,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:35,184 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:36,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:37,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:37,843 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:20:38,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:38,778 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:39,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:40,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:41,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:42,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:20:42,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:20:42,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:43,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:43,922 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:44,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:45,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:46,185 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:47,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:48,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:49,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:49,924 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:50,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:51,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:52,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:53,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:54,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:54,678 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:20:54,922 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:20:55,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:56,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:56,868 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:20:57,132 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:20:57,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:20:57,186 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:58,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:20:59,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:00,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:00,922 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:01,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:02,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:03,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:04,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:04,823 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:21:05,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:06,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:06,245 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:07,187 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:08,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:08,466 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:21:09,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:10,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:11,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:11,923 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:12,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:21:12,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:21:12,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:13,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:14,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:15,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:16,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:16,923 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:17,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:18,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:19,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:20,188 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:21,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:21,889 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:21:22,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:22,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:23,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:24,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:25,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:26,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:26,870 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:21:27,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:21:27,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:21:27,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:28,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:28,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:29,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:30,189 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:31,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:32,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:33,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:33,923 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:34,134 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:21:34,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:35,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:36,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:37,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:37,924 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:21:38,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:38,923 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:39,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:40,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:41,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:42,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:21:42,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:21:42,190 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:43,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:43,923 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:44,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:45,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:46,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:47,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:48,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:48,923 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:49,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:50,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:51,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:52,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:53,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:53,924 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:54,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:54,707 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:21:55,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:56,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:56,872 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:21:57,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:21:57,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:21:57,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:58,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:21:58,924 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:21:59,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:00,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:01,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:02,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:03,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:04,063 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:22:04,192 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:04,923 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:05,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:06,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:07,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:08,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:09,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:10,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:10,295 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:22:10,924 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:11,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:12,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:22:12,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:22:12,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:13,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:14,193 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:15,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:15,924 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:16,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:17,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:18,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:19,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:20,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:21,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:21,924 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:22,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:23,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:24,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:24,642 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:22:25,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:26,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:26,875 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:22:27,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:22:27,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:22:27,194 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:27,288 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:28,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:29,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:30,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:31,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:32,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:32,924 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:33,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:34,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:35,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:36,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:36,571 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:22:37,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:38,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:38,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:38,993 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:22:39,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:40,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:41,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:42,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:22:42,133 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:22:42,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:43,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:43,226 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:44,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:45,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:46,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:47,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:48,196 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:48,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:49,197 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:50,197 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:51,197 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:52,197 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:52,511 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:22:53,198 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:53,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:54,198 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:55,198 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:56,198 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:56,877 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:22:57,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:22:57,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:22:57,198 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:58,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:22:58,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:22:59,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:00,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:01,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:02,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:03,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:03,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:04,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:04,954 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:23:05,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:06,106 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:23:06,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:07,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:08,199 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:08,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:09,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:10,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:11,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:12,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:23:12,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:23:12,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:13,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:13,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:14,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:15,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:16,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:17,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:18,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:18,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:19,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:20,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:20,274 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:23:21,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:22,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:23,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:24,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:24,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:25,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:26,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:26,879 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:23:27,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:23:27,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:23:27,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:28,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:29,201 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:29,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:30,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:31,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:32,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:33,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:34,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:34,404 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:23:35,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:35,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:36,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:37,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:38,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:39,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:40,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:40,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:41,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:42,133 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:23:42,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:23:42,202 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:43,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:44,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:44,932 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:23:45,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:46,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:46,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:47,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:48,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:48,307 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:23:49,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:50,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:51,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:51,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:52,203 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:53,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:54,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:55,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:56,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:56,881 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:23:57,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:23:57,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:23:57,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:57,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:23:58,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:23:59,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:00,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:01,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:02,179 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:24:02,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:02,776 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:03,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:04,204 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:05,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:06,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:07,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:07,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:08,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:09,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:10,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:11,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:12,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:24:12,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:24:12,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:13,022 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:13,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:14,205 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:14,348 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:24:15,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:16,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:16,393 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:24:17,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:18,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:18,455 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:19,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:20,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:21,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:22,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:23,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:23,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:24,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:25,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:26,206 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:26,819 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:24:26,883 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:24:27,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:24:27,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:24:27,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:28,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:29,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:29,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:30,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:30,291 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:24:31,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:32,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:33,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:34,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:34,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:35,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:36,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:37,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:38,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:39,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:40,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:40,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:41,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:42,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:24:42,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:24:42,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:43,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:44,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:44,785 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:24:45,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:46,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:46,927 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:47,208 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:48,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:49,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:50,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:51,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:51,927 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:52,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:53,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:54,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:55,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:56,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:56,885 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:24:56,888 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:24:57,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:24:57,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:24:57,207 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:24:57,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:58,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:24:58,902 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:24:59,209 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:00,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:01,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:02,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:02,927 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:03,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:04,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:05,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:06,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:07,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:08,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:08,763 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:09,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:10,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:10,586 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:25:11,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:12,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:25:12,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:25:12,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:13,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:13,927 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:14,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:15,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:16,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:17,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:18,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:18,927 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:19,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:20,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:21,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:22,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:23,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:24,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:24,509 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:25:24,927 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:25,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:26,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:26,887 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:25:27,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:25:27,134 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:25:27,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:28,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:29,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:29,927 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:30,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:31,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:32,212 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:33,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:34,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:35,038 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:25:35,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:35,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:36,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:36,319 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:25:37,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:38,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:39,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:40,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:40,928 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:41,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:42,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:25:42,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:25:42,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:43,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:44,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:45,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:46,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:46,928 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:47,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:48,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:49,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:50,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:51,090 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:25:51,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:52,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:52,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:53,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:54,214 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:55,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:56,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:56,889 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:25:57,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:25:57,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:25:57,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:58,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:25:58,289 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:25:59,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:00,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:01,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:02,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:03,215 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:03,450 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:04,222 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:05,222 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:05,460 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:26:06,222 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:07,222 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:08,222 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:08,928 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:09,222 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:10,222 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:11,223 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:12,134 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:26:12,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:26:12,223 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:13,223 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:13,928 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:14,223 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:15,256 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:15,504 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:26:16,256 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:16,831 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:26:17,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:18,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:18,928 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:19,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:20,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:21,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:22,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:23,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:23,928 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:24,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:25,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:26,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:26,892 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:26:27,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:26:27,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:26:27,257 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:28,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:29,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:29,928 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:30,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:30,872 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:26:31,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:32,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:33,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:34,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:34,928 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:35,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:36,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:37,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:38,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:39,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:39,929 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:40,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:41,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:42,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:26:42,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:26:42,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:42,426 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:26:43,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:44,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:44,929 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:45,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:46,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:47,259 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:48,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:49,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:50,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:50,929 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:51,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:52,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:53,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:54,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:55,241 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:26:55,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:56,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:56,316 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:26:56,822 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:26:56,893 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:26:57,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:26:57,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:26:57,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:58,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:26:59,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:00,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:01,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:01,929 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:02,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:03,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:04,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:05,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:06,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:06,499 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:27:07,032 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:07,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:08,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:08,683 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:27:09,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:10,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:11,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:12,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:27:12,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:27:12,197 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:12,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:13,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:14,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:15,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:16,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:17,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:17,929 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:18,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:19,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:20,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:21,262 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:22,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:22,584 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:27:23,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:23,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:24,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:25,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:26,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:26,896 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:27:27,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:27:27,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:27:27,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:28,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:28,930 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:29,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:30,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:31,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:32,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:33,263 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:33,930 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:34,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:34,559 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:27:35,270 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:36,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:37,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:38,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:38,930 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:39,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:40,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:41,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:42,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:27:42,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:27:42,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:43,264 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:43,930 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:44,265 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:45,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:46,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:47,240 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:27:47,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:48,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:48,930 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:49,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:50,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:51,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:52,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:53,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:53,930 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:54,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:55,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:56,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:56,898 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:27:57,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:27:57,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:27:57,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:58,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:27:58,977 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:27:59,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:00,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:00,475 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:28:01,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:02,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:03,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:04,274 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:04,930 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:05,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:06,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:07,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:08,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:09,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:10,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:10,930 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:11,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:12,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:28:12,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:28:12,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:13,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:13,489 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:28:14,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:14,844 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:28:15,275 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:16,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:16,931 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:17,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:18,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:19,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:20,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:21,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:22,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:22,931 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:23,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:24,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:25,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:25,396 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:28:26,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:26,630 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:28:26,900 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:28:27,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:28:27,135 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:28:27,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:28,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:28,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:29,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:30,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:31,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:32,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:33,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:33,931 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:34,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:35,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:36,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:37,285 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:28:37,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:38,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:39,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:39,326 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:28:39,931 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:40,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:41,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:42,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:28:42,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:28:42,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:43,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:44,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:45,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:45,931 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:46,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:47,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:48,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:49,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:50,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:50,867 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:28:50,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:51,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:52,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:53,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:54,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:55,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:56,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:56,903 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:28:56,905 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:28:57,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:28:57,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:28:57,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:58,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:28:59,279 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:00,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:01,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:01,931 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:02,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:03,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:04,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:04,954 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:29:05,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:06,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:07,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:07,932 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:08,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:09,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:10,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:11,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:12,135 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:29:12,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:29:12,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:13,200 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:13,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:14,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:15,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:16,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:16,739 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:29:17,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:18,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:18,778 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:19,031 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:29:19,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:20,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:21,281 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:22,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:23,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:23,932 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:24,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:25,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:26,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:26,904 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:29:27,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:29:27,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:29:27,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:28,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:28,971 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:29,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:30,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:31,221 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:29:31,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:32,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:33,282 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:34,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:34,932 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:35,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:36,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:37,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:38,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:39,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:39,932 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:40,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:41,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:42,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:29:42,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:29:42,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:43,134 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:29:43,283 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:44,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:44,932 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:45,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:45,380 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:29:46,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:47,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:48,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:49,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:50,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:50,932 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:51,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:52,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:53,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:54,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:54,547 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:29:55,284 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:56,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:56,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:29:56,740 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:29:56,910 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:29:57,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:29:57,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:29:57,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:58,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:29:59,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:00,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:01,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:01,932 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:02,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:03,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:04,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:05,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:06,286 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:07,195 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:07,286 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:08,286 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:08,904 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:30:09,294 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:10,294 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:11,294 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:12,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:30:12,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:30:12,254 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:12,294 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:13,294 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:14,294 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:15,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:16,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:17,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:17,933 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:18,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:19,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:20,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:21,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:22,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:23,253 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:30:23,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:23,892 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:24,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:25,295 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:26,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:26,908 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:30:27,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:30:27,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:30:27,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:28,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:28,933 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:29,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:30,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:31,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:32,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:33,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:34,005 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:34,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:35,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:35,337 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:30:36,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:37,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:38,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:39,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:39,933 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:40,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:41,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:42,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:30:42,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:30:42,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:43,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:44,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:44,761 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:30:45,273 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:45,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:46,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:46,816 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:30:47,297 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:48,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:49,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:50,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:50,933 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:51,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:52,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:53,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:54,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:55,058 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:30:55,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:56,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:56,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:30:56,911 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:30:57,086 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:30:57,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:30:57,136 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:30:57,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:58,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:30:59,298 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:00,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:01,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:01,933 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:02,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:03,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:04,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:05,172 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:31:05,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:06,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:07,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:07,397 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:31:07,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:08,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:09,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:10,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:11,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:12,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:31:12,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:31:12,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:13,222 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:13,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:14,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:15,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:16,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:16,723 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:31:17,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:18,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:18,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:19,300 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:20,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:21,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:22,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:23,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:23,934 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:24,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:25,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:26,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:26,526 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:31:26,912 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:31:27,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:31:27,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:31:27,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:28,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:28,934 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:29,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:30,301 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:31,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:32,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:33,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:33,934 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:34,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:35,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:36,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:36,996 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:31:37,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:38,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:38,934 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:39,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:40,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:41,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:42,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:31:42,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:31:42,302 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:43,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:44,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:44,934 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:45,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:46,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:46,823 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:31:47,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:48,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:49,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:50,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:50,934 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:51,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:52,303 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:53,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:54,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:55,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:55,934 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:31:56,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:56,884 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:31:56,914 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:31:57,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:31:57,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:31:57,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:58,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:31:59,017 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:31:59,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:00,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:01,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:01,935 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:02,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:03,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:04,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:05,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:06,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:07,169 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:32:07,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:07,537 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:08,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:09,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:09,555 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:32:10,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:11,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:12,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:32:12,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:32:12,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:13,219 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:13,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:14,305 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:15,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:16,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:16,915 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:32:17,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:18,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:18,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:19,092 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:32:19,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:20,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:21,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:22,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:23,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:23,935 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:24,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:25,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:26,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:26,917 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:32:27,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:32:27,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:32:27,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:28,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:29,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:29,377 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:32:29,935 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:30,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:31,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:32,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:33,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:34,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:35,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:35,935 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:36,307 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:37,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:38,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:38,591 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:32:39,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:40,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:40,935 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:41,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:42,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:32:42,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:32:42,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:43,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:44,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:45,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:46,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:46,675 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:47,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:48,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:49,281 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:32:49,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:50,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:51,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:51,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:52,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:53,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:54,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:55,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:56,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:56,919 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:32:56,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:32:57,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:32:57,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:32:57,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:58,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:32:58,577 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:32:59,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:00,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:01,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:02,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:02,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:03,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:04,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:05,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:06,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:07,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:07,977 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:08,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:09,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:09,884 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:33:10,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:11,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:12,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:33:12,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:33:12,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:13,251 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:13,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:14,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:15,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:16,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:17,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:18,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:18,804 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:33:18,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:19,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:20,311 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:21,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:22,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:23,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:23,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:24,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:25,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:26,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:26,921 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:33:27,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:33:27,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:33:27,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:28,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:28,892 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:33:28,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:29,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:30,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:31,312 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:32,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:33,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:34,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:34,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:35,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:36,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:37,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:38,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:39,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:39,513 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:33:39,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:40,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:41,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:42,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:33:42,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:33:42,313 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:43,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:44,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:45,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:45,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:46,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:47,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:48,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:49,005 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:33:49,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:50,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:51,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:51,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:52,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:53,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:54,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:55,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:56,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:56,923 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:33:57,017 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:33:57,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:33:57,137 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:33:57,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:58,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:33:58,904 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:33:59,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:00,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:01,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:02,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:02,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:03,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:04,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:05,315 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:06,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:07,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:08,289 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:08,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:08,790 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:34:09,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:10,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:11,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:12,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:34:12,138 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:34:12,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:13,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:13,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:14,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:15,316 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:16,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:17,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:18,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:18,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:19,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:19,635 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:34:20,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:21,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:22,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:23,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:23,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:24,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:25,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:26,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:26,925 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:34:27,137 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:34:27,138 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:34:27,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:28,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:28,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:29,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:29,346 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:34:30,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:31,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:32,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:33,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:33,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:34,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:35,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:36,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:37,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:38,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:39,044 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:34:39,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:39,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:40,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:41,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:42,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:34:42,138 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:34:42,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:43,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:44,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:44,937 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:45,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:46,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:47,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:48,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:49,116 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:34:49,326 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:49,938 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:50,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:51,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:52,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:53,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:54,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:54,938 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:34:55,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:56,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:56,927 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:34:57,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:34:57,138 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:34:57,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:58,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:34:59,091 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:34:59,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:00,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:00,939 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:01,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:02,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:03,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:04,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:05,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:05,939 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:06,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:07,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:08,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:09,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:09,923 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:35:10,321 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:11,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:11,939 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:12,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:35:12,138 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:35:12,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:13,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:14,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:15,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:16,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:16,939 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:17,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:18,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:18,923 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:35:19,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:20,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:21,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:22,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:22,939 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:23,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:24,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:25,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:26,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:26,929 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:35:27,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:35:27,138 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:35:27,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:28,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:28,796 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:29,139 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:35:29,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:30,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:31,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:32,323 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:33,324 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:33,941 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:34,324 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:35,324 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:36,324 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:37,324 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:38,324 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:39,330 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:39,484 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:35:39,941 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:40,330 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:41,330 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:42,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:35:42,139 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:35:42,330 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:43,330 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:44,330 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:44,941 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:45,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:46,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:47,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:48,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:48,819 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:35:49,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:50,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:50,941 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:51,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:52,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:53,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:54,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:55,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:56,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:56,453 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:35:56,931 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:35:57,065 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:35:57,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:35:57,139 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:35:57,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:58,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:59,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:35:59,504 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:36:00,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:01,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:01,941 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:02,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:03,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:04,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:05,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:06,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:07,291 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:07,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:07,622 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:36:08,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:09,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:10,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:11,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:12,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:36:12,139 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:36:12,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:13,218 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:13,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:14,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:15,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:16,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:16,884 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:36:17,333 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:18,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:18,531 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:19,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:20,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:21,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:22,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:23,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:23,942 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:24,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:25,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:25,341 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:36:26,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:26,933 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:36:27,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:36:27,139 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:36:27,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:27,699 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:36:28,334 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:28,942 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:29,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:30,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:31,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:32,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:33,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:34,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:34,941 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:35,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:35,656 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:36:36,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:37,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:38,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:39,335 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:39,942 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:40,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:41,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:42,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:36:42,139 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:36:42,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:43,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:44,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:45,193 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:36:45,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:45,942 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:46,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:47,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:48,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:49,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:50,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:50,942 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:51,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:52,336 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:53,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:53,576 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:36:54,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:55,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:56,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:56,935 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:36:56,936 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:36:57,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:36:57,139 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:36:57,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:58,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:36:59,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:00,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:01,337 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:02,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:02,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:03,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:03,759 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:04,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:05,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:06,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:07,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:07,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:08,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:09,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:10,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:11,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:11,827 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:12,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:37:12,146 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:37:12,338 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:12,904 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:13,260 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:13,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:14,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:15,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:16,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:17,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:18,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:18,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:19,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:20,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:21,121 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:21,350 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:22,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:23,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:24,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:24,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:25,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:26,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:26,937 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:37:27,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:37:27,146 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:37:27,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:28,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:29,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:29,576 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:30,021 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:30,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:31,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:32,015 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:32,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:33,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:34,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:35,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:35,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:36,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:37,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:38,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:39,176 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:39,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:40,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:40,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:41,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:42,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:37:42,146 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:37:42,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:43,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:44,341 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:45,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:45,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:46,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:47,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:48,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:49,217 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:49,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:50,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:51,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:51,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:52,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:53,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:54,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:55,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:56,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:56,939 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:37:57,002 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:37:57,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:37:57,146 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:37:57,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:57,834 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:37:58,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:59,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:37:59,999 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:00,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:01,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:02,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:02,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:03,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:04,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:05,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:06,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:07,343 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:07,702 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:08,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:08,361 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:09,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:10,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:11,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:12,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:38:12,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:38:12,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:13,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:13,943 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:14,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:15,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:16,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:17,244 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:17,344 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:18,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:18,944 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:19,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:20,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:21,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:22,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:23,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:23,944 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:24,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:25,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:25,874 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:26,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:26,942 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:38:27,146 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:38:27,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:38:27,247 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:27,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:28,345 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:28,944 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:29,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:30,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:31,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:32,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:33,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:33,974 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:34,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:35,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:35,919 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:36,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:37,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:38,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:39,346 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:39,944 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:40,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:41,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:42,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:38:42,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:38:42,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:43,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:44,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:45,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:45,902 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:45,944 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:46,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:47,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:48,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:49,347 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:50,348 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:50,944 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:51,348 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:52,348 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:53,348 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:53,352 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:54,352 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:55,352 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:55,448 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:38:56,352 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:56,944 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:38:56,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:38:57,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:38:57,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:38:57,352 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:58,352 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:38:59,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:00,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:01,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:02,007 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:02,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:03,276 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:39:03,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:04,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:05,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:06,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:07,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:07,954 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:08,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:09,353 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:10,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:11,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:12,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:39:12,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:39:12,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:13,161 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:39:13,293 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:13,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:14,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:15,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:16,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:17,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:18,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:18,945 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:19,354 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:20,390 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:21,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:21,971 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:39:22,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:23,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:24,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:24,945 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:25,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:26,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:26,946 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:39:27,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:39:27,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:39:27,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:28,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:29,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:29,962 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:30,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:31,355 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:31,753 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:39:32,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:33,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:34,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:35,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:35,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:36,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:37,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:38,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:39,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:40,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:41,342 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:41,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:42,019 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:39:42,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:39:42,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:39:42,356 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:43,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:44,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:45,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:46,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:46,945 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:47,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:48,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:49,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:49,359 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:39:50,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:51,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:52,357 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:52,945 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:53,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:54,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:55,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:56,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:56,948 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:39:57,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:39:57,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:39:57,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:58,024 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:39:58,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:59,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:39:59,367 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:00,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:01,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:02,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:03,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:03,945 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:04,358 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:05,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:06,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:07,196 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:07,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:08,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:08,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:09,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:09,576 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:10,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:11,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:12,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:40:12,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:40:12,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:13,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:14,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:14,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:15,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:16,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:17,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:17,650 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:18,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:19,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:19,943 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:19,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:20,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:21,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:22,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:23,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:24,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:25,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:25,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:26,360 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:26,950 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:40:27,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:40:27,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:40:27,363 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:27,526 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:28,363 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:29,363 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:30,363 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:30,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:31,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:32,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:33,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:34,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:35,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:36,241 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:36,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:37,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:37,811 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:38,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:39,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:40,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:41,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:41,946 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:42,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:40:42,147 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:40:42,364 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:43,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:44,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:45,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:45,513 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:46,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:47,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:47,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:48,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:49,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:50,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:51,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:52,365 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:52,947 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:53,366 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:54,366 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:55,370 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:55,548 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:40:56,370 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:56,952 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:40:57,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:40:57,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:40:57,370 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:58,213 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:40:58,370 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:40:59,370 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:00,370 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:01,371 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:02,394 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:03,371 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:03,412 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:03,664 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:04,378 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:05,378 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:05,810 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:06,378 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:07,378 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:08,378 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:08,947 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:09,378 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:10,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:11,378 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:12,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:41:12,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:41:12,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:13,175 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:13,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:14,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:14,947 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:15,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:16,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:17,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:18,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:19,379 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:19,947 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:20,382 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:21,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:21,716 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:22,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:23,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:24,267 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:24,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:24,947 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:25,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:26,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:26,954 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:41:27,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:41:27,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:41:27,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:28,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:29,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:30,025 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:30,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:31,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:32,219 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:32,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:33,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:34,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:35,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:35,947 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:36,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:37,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:38,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:39,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:40,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:41,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:41,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:42,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:41:42,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:41:42,186 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:42,381 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:43,382 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:44,382 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:45,382 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:46,382 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:46,947 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:47,382 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:48,382 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:49,382 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:49,703 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:50,386 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:51,386 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:52,307 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:52,386 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:52,947 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:53,386 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:54,386 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:55,386 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:56,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:56,957 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:41:57,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:41:57,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:41:57,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:57,972 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:41:58,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:41:59,260 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:41:59,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:00,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:01,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:02,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:03,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:03,948 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:04,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:05,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:06,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:07,387 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:08,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:09,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:09,910 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:42:09,948 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:10,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:11,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:12,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:42:12,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:42:12,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:13,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:14,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:14,948 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:15,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:16,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:17,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:18,389 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:19,389 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:19,436 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:42:19,948 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:20,389 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:21,389 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:22,389 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:23,389 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:24,389 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:25,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:25,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:26,409 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:26,958 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:42:27,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:42:27,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:42:27,276 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:42:27,409 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:28,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:29,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:30,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:30,948 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:31,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:32,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:33,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:34,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:35,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:35,552 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:42:36,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:36,753 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:37,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:38,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:39,410 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:40,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:41,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:41,948 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:42,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:42:42,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:42:42,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:43,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:44,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:45,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:45,748 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:42:46,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:47,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:47,948 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:48,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:49,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:50,411 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:51,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:52,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:52,985 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:53,365 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:42:53,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:54,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:55,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:55,616 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:42:56,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:56,960 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:42:57,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:42:57,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:42:57,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:58,286 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:42:58,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:42:59,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:00,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:01,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:02,412 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:03,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:03,516 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:43:03,949 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:04,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:05,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:06,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:07,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:08,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:09,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:09,949 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:10,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:11,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:12,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:43:12,148 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:43:12,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:13,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:13,814 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:43:14,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:15,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:15,949 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:16,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:17,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:18,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:19,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:20,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:21,375 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:21,383 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:43:21,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:22,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:23,414 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:24,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:25,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:26,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:26,949 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:26,962 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:43:27,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:43:27,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:43:27,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:28,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:29,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:30,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:31,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:32,203 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:43:32,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:32,456 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:33,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:34,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:35,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:36,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:37,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:37,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:38,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:39,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:40,047 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:43:40,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:41,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:42,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:43:42,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:43:42,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:43,278 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:43,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:44,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:45,416 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:46,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:47,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:48,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:48,894 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:49,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:50,043 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:43:50,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:51,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:52,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:53,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:53,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:54,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:55,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:56,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:56,964 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:43:57,148 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:43:57,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:43:57,417 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:57,876 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:43:58,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:43:58,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:43:59,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:00,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:01,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:02,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:03,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:03,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:04,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:05,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:05,835 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:44:06,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:07,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:08,087 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:44:08,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:09,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:09,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:10,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:11,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:12,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:44:12,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:44:12,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:13,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:14,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:15,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:15,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:15,994 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:44:16,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:17,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:18,419 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:19,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:20,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:20,950 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:21,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:22,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:23,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:24,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:25,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:25,795 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:44:25,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:26,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:26,966 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:44:27,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:44:27,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:44:27,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:28,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:29,420 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:30,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:31,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:31,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:32,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:33,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:33,850 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:44:34,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:35,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:36,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:37,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:37,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:38,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:39,421 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:40,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:41,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:42,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:44:42,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:44:42,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:43,253 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:43,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:44,149 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:44:44,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:45,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:46,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:47,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:48,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:48,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:49,422 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:50,427 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:51,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:51,999 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:44:52,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:53,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:53,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:54,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:55,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:56,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:56,969 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:44:57,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:44:57,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:44:57,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:58,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:44:58,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:44:59,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:00,428 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:01,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:01,624 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:02,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:03,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:04,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:04,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:05,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:06,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:07,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:08,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:09,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:10,011 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:10,069 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:10,429 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:11,396 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:11,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:12,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:45:12,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:45:12,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:13,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:14,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:15,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:15,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:16,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:17,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:18,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:19,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:19,633 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:20,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:20,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:21,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:22,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:23,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:24,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:25,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:26,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:26,860 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:26,971 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:45:27,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:45:27,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:45:27,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:28,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:28,446 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:29,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:29,585 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:30,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:31,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:31,952 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:32,431 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:33,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:34,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:35,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:36,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:37,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:37,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:37,740 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:38,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:39,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:40,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:41,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:42,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:45:42,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:45:42,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:42,983 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:43,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:44,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:44,523 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:45,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:45,927 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:46,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:47,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:48,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:48,952 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:49,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:50,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:51,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:51,739 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:52,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:53,433 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:54,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:54,952 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:45:55,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:56,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:56,973 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:45:57,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:45:57,149 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:45:57,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:58,251 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:45:58,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:59,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:45:59,952 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:00,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:00,511 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:01,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:02,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:03,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:04,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:05,029 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:05,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:06,067 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:06,434 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:07,412 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:07,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:08,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:09,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:10,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:10,952 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:11,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:12,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:46:12,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:46:12,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:13,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:14,055 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:14,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:15,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:16,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:16,952 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:17,435 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:18,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:19,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:20,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:21,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:21,578 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:22,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:22,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:23,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:24,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:25,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:26,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:26,975 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:46:27,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:46:27,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:46:27,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:27,790 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:28,436 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:29,437 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:29,692 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:30,437 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:31,437 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:32,437 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:32,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:33,437 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:34,437 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:35,437 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:36,437 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:37,442 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:38,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:38,539 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:38,864 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:39,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:40,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:41,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:42,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:46:42,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:46:42,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:43,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:43,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:44,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:45,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:46,395 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:46,438 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:47,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:48,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:48,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:49,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:50,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:51,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:52,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:52,463 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:53,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:53,691 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:46:53,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:54,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:55,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:56,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:56,977 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:46:57,149 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:46:57,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:46:57,439 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:58,440 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:46:59,020 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:46:59,440 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:00,440 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:00,544 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:01,440 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:02,440 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:03,440 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:04,440 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:04,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:05,440 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:06,457 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:07,457 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:08,267 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:08,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:09,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:09,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:10,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:11,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:12,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:47:12,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:47:12,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:13,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:14,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:15,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:15,808 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:16,067 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:16,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:17,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:18,458 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:19,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:20,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:20,953 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:21,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:22,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:23,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:23,830 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:24,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:25,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:26,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:26,954 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:26,979 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:47:27,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:47:27,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:47:27,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:28,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:29,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:30,459 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:31,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:32,325 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:32,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:32,587 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:33,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:34,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:35,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:36,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:37,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:37,896 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:38,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:39,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:40,460 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:40,539 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:41,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:42,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:47:42,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:47:42,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:43,230 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:43,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:44,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:45,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:46,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:47,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:48,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:48,543 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:48,924 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:49,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:50,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:51,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:52,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:53,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:53,842 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:53,954 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:54,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:55,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:56,178 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:47:56,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:56,981 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:47:57,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:47:57,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:47:57,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:58,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:47:58,954 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:47:59,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:00,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:01,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:02,031 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:02,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:03,462 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:03,954 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:04,131 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:04,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:05,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:06,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:07,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:08,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:08,954 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:09,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:10,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:10,740 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:11,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:12,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:48:12,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:48:12,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:13,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:14,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:14,954 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:15,463 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:16,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:17,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:18,210 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:18,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:19,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:20,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:20,954 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:21,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:22,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:23,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:24,437 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:24,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:25,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:26,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:26,955 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:26,984 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:48:27,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:48:27,150 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:48:27,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:28,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:29,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:30,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:31,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:31,665 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:32,265 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:32,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:33,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:34,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:35,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:36,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:37,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:37,465 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:38,330 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:38,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:39,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:40,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:40,595 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:41,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:42,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:48:42,151 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:48:42,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:43,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:43,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:44,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:45,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:45,779 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:46,466 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:47,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:48,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:48,784 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:49,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:50,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:51,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:52,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:53,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:53,955 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:54,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:54,519 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:48:55,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:56,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:56,986 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:48:57,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:48:57,151 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:48:57,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:58,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:59,037 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:48:59,467 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:48:59,549 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:00,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:01,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:02,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:03,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:04,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:04,955 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:05,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:06,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:07,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:08,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:08,505 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:09,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:09,955 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:10,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:11,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:12,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:49:12,151 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:49:12,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:13,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:14,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:14,745 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:15,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:15,939 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:15,956 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:16,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:17,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:18,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:19,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:20,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:21,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:21,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:21,763 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:22,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:23,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:24,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:25,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:26,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:26,956 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:26,988 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:49:27,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:49:27,151 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:49:27,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:27,683 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:28,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:29,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:29,967 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:30,470 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:31,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:32,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:32,956 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:33,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:34,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:35,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:35,923 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:36,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:37,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:38,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:38,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:39,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:40,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:41,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:42,150 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:49:42,151 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:49:42,471 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:43,270 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:43,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:44,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:44,787 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:45,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:46,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:47,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:48,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:48,956 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:49,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:49,907 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:50,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:51,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:52,015 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:52,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:53,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:53,956 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:54,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:55,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:56,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:56,990 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:49:57,151 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:49:57,151 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:49:57,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:57,923 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:49:58,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:49:58,956 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:49:59,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:00,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:01,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:02,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:03,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:04,004 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:04,473 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:05,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:05,794 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:06,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:07,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:08,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:09,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:09,956 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:10,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:11,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:11,807 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:12,151 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:50:12,151 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:50:12,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:13,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:14,474 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:15,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:15,957 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:16,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:17,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:18,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:19,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:19,854 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:20,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:20,957 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:21,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:22,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:23,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:24,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:25,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:26,007 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:26,461 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:26,475 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:26,996 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:50:27,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:50:27,153 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:50:27,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:28,027 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:28,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:29,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:30,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:31,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:31,957 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:32,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:33,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:33,827 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:34,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:35,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:36,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:37,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:37,957 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:38,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:39,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:40,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:41,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:41,724 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:42,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:50:42,153 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:50:42,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:43,285 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:43,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:44,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:45,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:46,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:47,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:48,404 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:48,477 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:48,687 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:49,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:50,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:51,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:52,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:53,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:53,957 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:54,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:55,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:55,819 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:50:56,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:56,998 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:50:57,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:50:57,153 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:50:57,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:58,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:50:58,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:50:59,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:00,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:01,478 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:02,309 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:02,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:03,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:03,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:04,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:04,707 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:05,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:06,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:07,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:08,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:09,017 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:09,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:09,942 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:10,479 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:11,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:12,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:51:12,153 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:51:12,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:13,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:14,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:14,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:15,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:16,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:17,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:18,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:18,867 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:19,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:20,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:20,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:21,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:22,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:23,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:23,851 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:24,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:25,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:26,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:26,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:27,000 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:51:27,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:51:27,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:51:27,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:28,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:29,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:30,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:31,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:32,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:32,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:32,499 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:33,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:34,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:35,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:36,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:37,191 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:37,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:38,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:38,693 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:39,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:40,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:41,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:42,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:51:42,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:51:42,242 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:42,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:43,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:44,482 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:45,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:45,990 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:46,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:47,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:47,517 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:48,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:49,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:50,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:51,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:52,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:52,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:52,779 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:53,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:53,875 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:51:54,483 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:55,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:56,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:57,002 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:51:57,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:51:57,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:51:57,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:58,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:51:58,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:51:59,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:00,182 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:00,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:01,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:02,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:03,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:03,959 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:04,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:05,484 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:06,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:06,746 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:07,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:08,010 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:08,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:08,959 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:09,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:10,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:11,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:12,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:52:12,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:52:12,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:13,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:13,959 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:14,390 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:14,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:15,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:16,485 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:17,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:18,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:18,959 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:19,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:20,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:21,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:22,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:22,747 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:23,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:23,959 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:24,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:25,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:26,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:27,004 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:52:27,153 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:52:27,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:52:27,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:28,486 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:28,579 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:28,963 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:29,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:30,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:31,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:32,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:33,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:34,138 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:34,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:35,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:36,148 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:36,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:37,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:38,487 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:39,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:39,959 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:40,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:41,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:42,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:52:42,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:52:42,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:43,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:44,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:44,959 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:45,020 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:45,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:46,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:47,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:48,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:49,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:50,489 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:50,651 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:50,752 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:51,489 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:52,031 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:52,489 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:53,489 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:54,489 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:55,489 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:55,960 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:52:56,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:57,006 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:52:57,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:52:57,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:52:57,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:58,183 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:52:58,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:52:59,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:00,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:01,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:01,960 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:02,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:03,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:04,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:05,521 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:06,143 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:06,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:07,160 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:07,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:08,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:09,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:10,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:11,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:12,016 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:12,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:53:12,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:53:12,293 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:12,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:13,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:14,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:15,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:16,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:17,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:17,960 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:18,522 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:19,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:19,823 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:20,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:21,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:22,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:23,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:23,960 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:24,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:25,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:26,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:26,807 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:27,008 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:53:27,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:53:27,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:53:27,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:28,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:28,960 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:29,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:30,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:31,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:32,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:33,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:34,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:34,936 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:34,960 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:35,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:36,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:37,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:38,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:39,524 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:40,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:40,707 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:40,791 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:41,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:41,923 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:42,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:53:42,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:53:42,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:43,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:44,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:45,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:45,961 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:46,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:47,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:48,170 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:48,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:49,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:50,525 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:51,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:51,961 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:52,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:53,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:54,003 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:53:54,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:55,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:56,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:57,010 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:53:57,013 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:53:57,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:53:57,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:53:57,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:58,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:53:59,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:00,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:01,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:02,158 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:02,526 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:02,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:03,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:04,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:05,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:06,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:07,530 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:07,796 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:08,267 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:08,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:09,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:10,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:10,567 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:11,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:12,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:54:12,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:54:12,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:13,277 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:13,527 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:14,528 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:15,528 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:16,373 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:16,530 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:17,528 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:18,528 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:18,961 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:19,528 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:20,531 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:21,531 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:22,215 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:22,534 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:23,534 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:23,961 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:24,536 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:24,603 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:25,537 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:26,539 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:27,012 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:54:27,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:54:27,154 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:54:27,539 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:28,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:29,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:29,961 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:30,326 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:30,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:31,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:32,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:33,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:34,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:34,961 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:35,574 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:36,223 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:36,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:37,540 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:38,427 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:38,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:39,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:39,961 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:40,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:41,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:42,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:54:42,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:54:42,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:43,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:44,131 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:44,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:45,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:45,961 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:46,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:47,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:48,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:49,541 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:49,966 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:50,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:51,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:51,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:52,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:53,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:54,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:55,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:56,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:57,014 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:54:57,017 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:54:57,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:54:57,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:54:57,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:58,244 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:54:58,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:54:59,542 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:00,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:01,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:02,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:02,962 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:03,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:03,999 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:55:04,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:05,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:06,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:07,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:08,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:08,962 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:09,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:10,543 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:11,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:12,006 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:55:12,154 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:55:12,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:55:12,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:13,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:13,962 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:14,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:15,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:16,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:17,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:18,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:18,815 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:55:18,962 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:19,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:20,544 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:21,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:22,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:23,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:24,041 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:24,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:25,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:26,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:26,859 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:55:27,016 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:55:27,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:55:27,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:55:27,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:28,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:29,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:29,962 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:30,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:31,545 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:32,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:32,779 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:55:33,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:34,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:35,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:35,962 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:36,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:37,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:38,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:39,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:40,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:41,053 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:55:41,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:41,963 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:42,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:55:42,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:55:42,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:43,546 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:44,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:45,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:46,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:46,759 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:55:47,504 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:47,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:48,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:49,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:50,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:51,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:52,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:53,413 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:53,547 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:54,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:54,928 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:55:55,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:56,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:57,019 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:55:57,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:55:57,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:55:57,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:58,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:55:58,926 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:55:59,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:00,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:00,744 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:01,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:02,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:03,154 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:03,548 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:03,963 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:04,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:05,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:06,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:07,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:08,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:08,665 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:09,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:09,963 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:10,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:11,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:12,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:56:12,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:56:12,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:13,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:14,531 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:14,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:14,972 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:15,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:16,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:16,747 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:17,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:18,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:19,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:20,469 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:20,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:21,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:22,292 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:22,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:23,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:24,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:25,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:25,963 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:26,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:27,021 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:56:27,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:56:27,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:56:27,550 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:28,443 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:28,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:29,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:30,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:31,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:31,963 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:32,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:33,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:34,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:35,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:36,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:36,795 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:37,513 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:37,551 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:38,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:39,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:40,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:41,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:42,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:56:42,155 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:56:42,435 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:42,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:43,266 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:43,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:44,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:45,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:46,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:47,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:48,320 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:48,552 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:49,161 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:49,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:50,211 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:50,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:51,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:52,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:53,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:53,964 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:54,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:55,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:56,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:56,938 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:56:57,023 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:56:57,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:56:57,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:56:57,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:58,553 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:56:58,964 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:56:59,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:00,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:01,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:02,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:03,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:03,964 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:04,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:05,076 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:05,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:06,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:07,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:08,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:08,964 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:09,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:10,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:10,710 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:11,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:12,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:57:12,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:57:12,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:13,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:14,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:14,964 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:15,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:16,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:17,068 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:17,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:18,077 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:18,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:19,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:19,964 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:20,555 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:21,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:22,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:23,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:24,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:25,093 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:25,430 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:25,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:26,129 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:26,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:27,025 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:57:27,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:57:27,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:57:27,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:28,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:29,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:30,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:30,965 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:31,556 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:32,174 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:32,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:33,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:34,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:35,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:36,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:36,965 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:37,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:38,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:39,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:40,210 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:40,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:41,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:41,965 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:42,155 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:57:42,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:57:42,557 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:43,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:44,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:45,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:46,334 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:46,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:47,326 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:47,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:48,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:49,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:50,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:51,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:52,283 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:52,532 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:52,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:53,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:54,552 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:57:54,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:55,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:56,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:57,027 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:57:57,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:57:57,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:57:57,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:58,232 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:57:58,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:57:59,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:00,240 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:00,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:01,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:02,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:03,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:03,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:04,559 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:05,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:06,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:06,899 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:07,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:08,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:08,965 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:09,088 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:09,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:10,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:11,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:12,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:58:12,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:58:12,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:13,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:14,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:14,965 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:15,399 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:15,560 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:16,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:17,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:18,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:19,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:20,013 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:20,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:20,887 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:21,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:22,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:23,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:24,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:25,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:25,966 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:26,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:27,029 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:58:27,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:58:27,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:58:27,561 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:28,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:28,980 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:29,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:30,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:31,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:31,966 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:32,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:33,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:34,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:34,611 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:35,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:36,562 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:37,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:37,966 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:38,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:39,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:40,203 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:40,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:41,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:42,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:58:42,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:58:42,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:42,998 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:43,296 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:43,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:44,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:45,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:46,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:47,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:48,563 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:48,611 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:48,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:49,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:50,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:51,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:52,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:53,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:53,966 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:58:54,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:55,218 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:55,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:56,230 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:58:56,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:57,031 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:58:57,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:58:57,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:58:57,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:58,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:59,564 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:58:59,966 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:00,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:01,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:02,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:02,974 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:03,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:04,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:05,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:05,966 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:06,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:07,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:08,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:09,028 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:09,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:10,355 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:10,565 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:11,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:11,966 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:12,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:59:12,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:59:12,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:13,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:14,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:15,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:16,215 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:16,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:17,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:17,929 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:18,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:19,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:20,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:21,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:22,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:23,141 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:23,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:23,578 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:24,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:25,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:26,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:27,033 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:59:27,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:59:27,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:59:27,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:28,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:28,665 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:29,153 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:29,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:30,255 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:30,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:31,567 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:32,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:33,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:33,967 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:34,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:35,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:36,423 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:36,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:37,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:38,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:38,967 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:39,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:40,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:41,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:42,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:59:42,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:59:42,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:43,548 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:43,568 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:43,967 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:44,569 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:44,795 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:45,569 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:46,569 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:47,569 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:48,579 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:49,579 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:49,967 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:50,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:50,597 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:51,579 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:52,579 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:53,579 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:54,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:54,967 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 11:59:55,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:56,278 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 11:59:56,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:57,035 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 11:59:57,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 11:59:57,156 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 11:59:57,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:58,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 11:59:59,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:00,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:00,967 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:01,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:02,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:03,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:04,395 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:04,580 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:05,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:06,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:06,967 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:07,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:08,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:09,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:10,271 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:10,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:11,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:12,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:00:12,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:00:12,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:12,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:12,903 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:13,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:14,581 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:15,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:16,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:17,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:17,986 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:18,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:19,347 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:19,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:20,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:21,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:22,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:23,454 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:23,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:24,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:25,090 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:25,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:26,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:27,037 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:00:27,156 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:00:27,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:00:27,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:28,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:28,968 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:29,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:30,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:31,161 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:31,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:32,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:33,493 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:33,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:33,968 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:34,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:35,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:36,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:37,583 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:38,260 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:38,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:38,968 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:39,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:40,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:41,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:42,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:00:42,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:00:42,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:43,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:43,993 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:44,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:45,100 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:45,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:46,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:47,221 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:47,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:48,584 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:49,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:49,968 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:50,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:51,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:52,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:52,793 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:53,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:54,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:54,968 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:00:55,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:56,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:57,039 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:00:57,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:00:57,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:00:57,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:58,585 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:00:59,112 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:00:59,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:00,332 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:00,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:00,968 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:01,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:02,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:03,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:04,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:04,973 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:05,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:06,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:06,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:07,156 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:07,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:08,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:09,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:10,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:11,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:11,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:12,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:01:12,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:01:12,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:12,660 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:13,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:14,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:14,859 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:15,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:16,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:16,969 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:17,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:18,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:19,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:20,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:20,762 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:21,587 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:22,366 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:22,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:23,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:24,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:25,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:26,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:27,041 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:01:27,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:01:27,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:01:27,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:28,254 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:28,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:29,237 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:29,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:30,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:31,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:32,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:33,318 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:33,588 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:34,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:34,591 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:35,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:36,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:37,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:38,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:38,969 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:39,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:40,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:41,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:42,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:01:42,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:01:42,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:42,912 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:43,589 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:43,969 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:44,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:45,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:46,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:47,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:48,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:48,969 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:49,005 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:49,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:50,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:51,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:52,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:53,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:53,992 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:01:54,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:54,735 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:55,590 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:56,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:57,043 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:01:57,096 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:01:57,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:01:57,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:01:57,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:58,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:59,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:01:59,969 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:00,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:01,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:02,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:03,148 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:03,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:04,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:05,591 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:05,969 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:06,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:07,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:08,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:08,940 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:09,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:10,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:11,096 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:11,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:11,970 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:12,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:02:12,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:02:12,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:13,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:14,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:15,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:16,523 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:16,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:17,586 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:17,592 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:18,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:19,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:20,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:21,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:22,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:23,140 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:23,272 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:23,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:24,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:25,519 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:25,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:26,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:27,045 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:02:27,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:02:27,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:02:27,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:28,234 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:28,593 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:29,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:30,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:31,148 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:31,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:32,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:33,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:33,970 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:34,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:35,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:36,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:36,932 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:37,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:38,594 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:39,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:39,970 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:40,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:41,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:42,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:02:42,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:02:42,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:42,931 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:43,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:44,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:45,213 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:45,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:45,970 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:46,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:47,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:48,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:49,595 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:50,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:50,985 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:51,399 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:51,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:52,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:53,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:54,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:55,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:55,995 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:02:56,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:57,047 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:02:57,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:02:57,157 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:02:57,461 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:57,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:58,575 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:02:58,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:02:59,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:00,596 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:01,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:01,971 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:02,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:03,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:04,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:05,313 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:05,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:06,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:07,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:07,971 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:08,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:09,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:10,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:11,012 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:11,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:12,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:03:12,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:03:12,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:13,239 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:13,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:14,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:15,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:16,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:17,300 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:17,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:18,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:18,915 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:19,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:20,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:21,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:22,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:23,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:23,971 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:24,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:24,655 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:25,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:26,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:27,050 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:03:27,157 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:03:27,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:03:27,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:28,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:28,989 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:29,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:30,534 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:30,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:31,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:32,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:33,599 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:34,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:34,971 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:35,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:36,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:36,667 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:37,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:38,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:38,928 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:39,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:40,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:40,971 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:41,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:42,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:03:42,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:03:42,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:43,600 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:44,601 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:45,313 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:45,601 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:46,602 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:46,971 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:47,602 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:48,603 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:49,603 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:50,604 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:51,305 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:51,604 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:52,053 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:52,604 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:53,604 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:54,604 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:55,604 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:56,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:56,920 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:03:57,052 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:03:57,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:03:57,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:03:57,317 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:03:57,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:58,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:03:59,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:00,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:01,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:02,340 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:02,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:03,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:04,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:05,301 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:05,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:06,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:07,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:07,972 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:08,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:09,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:10,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:10,986 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:11,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:12,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:04:12,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:04:12,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:13,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:13,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:14,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:15,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:16,587 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:16,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:17,606 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:18,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:18,863 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:18,972 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:19,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:20,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:21,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:22,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:23,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:24,518 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:24,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:24,972 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:25,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:26,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:27,054 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:04:27,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:04:27,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:04:27,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:28,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:29,607 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:30,476 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:30,563 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:30,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:31,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:32,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:33,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:34,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:35,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:35,839 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:36,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:36,642 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:37,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:38,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:38,956 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:39,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:40,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:40,972 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:41,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:42,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:04:42,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:04:42,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:43,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:44,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:44,989 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:45,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:45,972 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:46,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:47,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:48,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:49,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:50,609 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:51,057 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:51,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:51,927 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:52,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:53,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:54,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:55,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:56,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:57,056 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:04:57,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:04:57,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:04:57,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:04:57,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:58,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:04:59,371 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:04:59,610 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:00,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:01,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:02,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:02,973 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:03,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:04,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:04,928 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:05,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:06,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:07,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:08,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:08,973 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:09,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:10,611 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:10,932 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:11,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:12,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:05:12,159 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:05:12,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:13,046 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:13,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:13,973 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:14,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:15,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:16,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:17,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:18,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:19,172 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:19,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:19,973 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:20,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:21,428 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:21,612 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:22,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:23,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:24,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:25,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:25,933 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:26,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:27,058 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:05:27,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:05:27,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:05:27,420 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:27,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:28,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:29,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:30,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:30,973 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:31,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:32,613 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:32,938 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:33,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:34,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:35,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:36,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:36,973 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:37,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:38,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:39,004 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:39,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:40,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:41,352 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:41,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:42,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:05:42,159 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:05:42,322 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:42,614 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:43,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:44,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:45,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:46,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:47,275 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:47,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:47,873 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:48,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:49,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:50,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:51,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:52,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:52,843 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:53,576 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:53,615 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:54,616 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:55,616 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:56,616 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:57,060 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:05:57,158 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:05:57,158 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:05:57,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:58,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:05:58,728 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:05:58,974 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:05:59,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:00,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:01,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:02,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:03,618 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:04,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:04,974 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:05,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:05,843 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:06,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:07,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:08,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:09,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:09,974 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:10,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:11,324 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:11,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:12,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:06:12,159 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:06:12,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:13,619 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:14,620 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:15,130 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:15,391 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:15,620 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:16,620 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:17,620 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:18,620 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:19,620 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:20,480 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:20,620 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:21,002 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:21,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:22,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:23,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:24,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:25,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:26,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:26,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:26,778 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:27,062 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:06:27,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:06:27,159 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:06:27,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:28,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:29,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:30,630 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:31,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:32,326 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:32,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:33,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:33,641 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:34,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:35,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:36,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:37,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:38,116 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:38,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:39,253 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:39,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:40,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:41,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:42,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:06:42,159 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:06:42,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:43,622 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:43,759 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:44,623 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:44,916 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:45,623 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:46,623 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:47,623 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:48,623 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:48,974 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:49,516 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:49,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:50,630 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:51,630 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:52,630 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:53,630 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:54,167 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:54,630 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:55,604 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:55,630 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:56,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:56,904 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:06:57,064 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:06:57,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:06:57,159 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:06:57,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:58,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:06:59,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:06:59,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:00,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:01,467 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:01,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:02,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:03,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:04,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:04,975 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:05,631 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:06,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:07,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:07,723 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:08,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:09,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:09,975 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:10,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:11,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:12,159 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:07:12,159 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:07:12,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:13,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:13,659 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:14,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:15,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:15,814 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:16,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:17,568 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:17,632 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:18,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:18,775 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:19,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:20,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:20,975 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:21,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:22,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:23,575 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:23,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:24,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:25,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:25,975 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:26,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:27,066 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:07:27,162 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:07:27,162 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:07:27,633 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:28,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:29,196 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:29,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:30,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:31,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:31,975 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:32,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:33,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:34,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:34,710 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:35,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:36,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:36,975 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:37,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:38,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:39,468 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:39,634 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:40,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:41,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:41,843 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:42,162 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:07:42,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:07:42,227 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:42,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:43,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:44,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:45,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:46,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:46,863 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:47,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:47,776 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:48,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:49,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:50,635 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:51,507 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:51,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:52,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:53,136 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:53,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:54,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:55,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:56,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:57,068 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:07:57,162 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:07:57,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:07:57,240 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:07:57,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:58,271 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:07:58,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:07:59,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:00,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:01,636 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:02,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:03,019 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:03,304 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:03,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:04,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:05,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:06,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:07,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:08,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:08,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:08,928 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:09,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:10,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:11,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:12,162 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:08:12,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:08:12,637 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:13,393 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:13,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:13,976 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:14,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:15,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:16,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:17,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:18,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:19,269 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:19,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:19,976 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:20,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:21,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:22,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:23,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:24,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:24,799 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:24,976 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:25,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:26,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:27,070 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:08:27,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:08:27,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:08:27,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:28,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:29,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:30,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:30,900 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:30,976 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:31,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:32,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:33,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:34,639 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:35,557 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:35,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:36,219 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:36,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:37,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:38,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:39,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:40,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:41,472 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:41,557 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:41,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:42,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:08:42,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:08:42,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:43,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:44,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:45,640 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:46,645 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:46,870 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:47,645 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:47,699 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:48,645 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:49,645 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:50,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:51,332 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:51,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:52,085 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:52,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:53,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:54,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:55,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:56,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:57,072 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:08:57,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:08:57,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:08:57,299 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:08:57,567 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:08:57,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:58,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:08:59,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:00,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:01,012 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:01,646 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:02,626 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:02,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:03,413 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:03,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:04,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:05,686 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:06,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:06,823 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:07,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:07,951 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:08,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:09,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:10,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:11,647 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:12,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:09:12,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:09:12,666 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:13,276 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:13,647 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:13,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:14,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:15,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:16,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:17,490 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:17,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:18,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:18,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:18,835 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:19,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:20,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:21,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:22,648 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:23,649 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:23,977 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:24,649 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:25,649 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:25,735 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:26,649 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:27,074 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:09:27,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:09:27,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:09:27,649 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:28,649 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:28,977 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:29,487 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:29,649 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:30,649 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:31,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:31,891 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:32,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:33,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:34,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:34,977 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:35,349 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:35,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:36,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:37,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:38,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:39,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:40,168 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:40,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:41,401 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:41,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:42,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:09:42,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:09:42,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:43,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:44,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:45,144 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:45,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:45,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:46,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:47,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:48,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:49,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:50,508 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:50,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:50,992 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:51,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:52,651 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:53,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:54,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:55,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:55,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:09:55,699 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:56,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:56,916 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:09:57,076 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:09:57,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:09:57,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:09:57,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:58,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:09:59,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:00,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:00,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:01,325 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:01,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:02,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:03,652 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:04,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:05,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:06,504 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:06,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:07,181 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:07,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:08,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:09,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:10,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:11,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:11,715 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:11,884 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:12,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:10:12,163 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:10:12,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:13,427 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:13,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:14,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:15,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:16,653 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:17,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:17,484 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:17,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:18,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:19,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:19,802 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:20,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:21,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:22,286 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:22,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:23,451 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:23,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:24,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:25,654 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:26,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:27,079 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:10:27,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:10:27,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:10:27,332 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:27,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:28,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:29,320 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:29,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:30,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:31,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:32,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:32,815 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:33,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:34,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:35,497 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:35,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:36,655 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:37,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:37,979 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:38,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:39,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:40,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:41,412 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:41,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:42,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:10:42,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:10:42,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:43,289 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:43,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:44,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:45,407 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:45,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:46,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:47,656 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:47,924 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:48,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:48,664 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:49,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:50,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:51,340 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:51,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:52,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:53,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:53,812 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:54,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:55,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:56,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:57,080 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:10:57,163 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:10:57,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:10:57,436 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:10:57,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:58,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:10:58,979 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:10:59,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:00,657 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:01,172 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:01,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:02,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:03,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:03,979 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:04,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:05,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:06,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:07,309 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:07,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:08,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:08,979 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:09,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:10,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:11,658 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:12,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:11:12,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:11:12,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:13,255 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:13,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:13,979 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:14,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:15,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:16,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:17,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:18,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:18,979 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:19,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:20,072 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:20,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:21,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:22,659 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:23,509 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:23,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:24,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:24,979 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:25,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:26,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:27,083 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:11:27,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:11:27,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:11:27,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:28,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:29,615 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:29,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:30,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:30,979 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:31,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:32,660 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:33,233 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:33,661 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:34,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:35,661 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:35,701 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:36,661 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:36,980 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:37,661 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:38,661 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:39,293 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:39,661 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:40,661 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:41,470 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:41,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:42,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:11:42,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:11:42,251 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:42,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:43,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:44,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:45,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:46,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:47,554 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:47,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:47,767 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:48,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:49,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:50,662 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:51,293 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:51,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:52,605 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:52,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:53,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:54,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:55,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:56,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:57,026 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:11:57,085 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:11:57,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:11:57,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:11:57,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:58,249 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:11:58,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:11:59,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:00,664 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:01,663 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:02,664 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:03,666 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:03,902 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:04,227 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:04,666 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:05,666 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:06,666 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:07,666 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:08,666 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:08,985 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:09,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:10,099 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:10,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:11,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:12,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:12:12,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:12:12,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:13,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:13,822 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:13,985 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:14,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:15,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:16,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:17,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:18,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:19,029 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:19,592 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:19,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:20,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:21,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:22,003 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:22,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:23,160 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:23,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:24,082 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:24,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:25,522 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:25,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:26,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:27,087 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:12:27,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:12:27,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:12:27,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:28,187 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:28,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:29,122 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:29,344 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:29,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:30,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:31,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:31,760 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:32,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:33,029 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:33,668 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:34,228 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:34,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:35,248 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:35,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:36,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:37,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:37,823 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:38,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:39,319 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:39,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:40,202 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:40,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:41,487 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:41,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:42,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:12:42,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:12:42,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:43,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:43,922 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:44,398 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:44,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:45,336 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:45,669 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:46,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:47,532 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:47,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:48,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:49,432 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:49,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:49,931 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:50,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:51,140 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:51,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:52,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:53,595 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:53,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:54,468 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:54,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:55,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:56,004 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:56,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:57,089 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:12:57,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:12:57,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:12:57,221 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:57,670 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:58,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:12:59,469 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:12:59,532 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:12:59,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:00,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:01,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:01,739 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:02,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:03,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:03,943 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:04,558 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:04,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:05,053 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:05,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:06,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:07,328 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:07,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:08,671 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:09,566 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:09,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:09,908 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:10,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:11,092 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:11,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:12,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:13:12,164 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:13:12,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:13,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:13,776 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:14,608 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:14,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:15,077 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:15,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:16,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:17,461 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:17,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:18,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:19,672 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:19,683 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:19,986 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:20,673 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:21,673 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:22,673 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:23,127 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:23,673 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:24,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:24,986 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:25,281 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:25,686 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:26,686 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:27,091 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:13:27,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:13:27,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:13:27,686 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:28,686 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:29,686 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:30,032 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:30,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:30,986 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:31,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:32,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:33,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:34,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:35,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:35,992 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:36,258 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:36,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:37,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:38,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:39,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:40,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:41,280 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:41,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:41,763 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:42,164 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:13:42,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:13:42,687 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:43,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:44,178 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:44,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:45,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:46,326 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:46,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:47,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:47,851 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:48,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:49,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:50,143 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:50,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:51,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:51,987 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:52,688 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:53,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:53,689 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:54,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:55,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:56,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:56,987 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:13:57,093 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:13:57,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:13:57,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:13:57,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:58,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:13:59,195 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:13:59,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:00,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:01,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:02,628 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:02,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:03,689 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:03,890 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:04,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:05,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:06,227 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:06,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:07,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:07,998 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:08,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:09,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:10,051 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:10,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:11,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:12,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:14:12,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:14:12,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:13,239 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:13,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:14,690 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:15,373 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:15,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:16,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:17,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:18,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:18,987 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:19,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:20,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:21,344 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:21,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:22,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:23,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:23,987 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:24,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:25,619 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:25,691 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:26,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:27,095 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:14:27,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:14:27,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:14:27,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:28,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:28,987 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:29,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:30,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:31,581 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:31,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:32,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:33,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:33,987 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:34,063 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:34,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:35,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:36,692 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:37,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:37,711 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:38,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:38,987 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:39,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:40,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:41,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:42,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:14:42,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:14:42,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:43,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:43,987 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:44,074 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:44,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:45,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:46,693 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:47,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:47,699 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:48,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:48,988 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:49,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:50,071 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:50,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:51,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:52,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:53,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:53,949 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:14:53,988 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:54,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:55,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:56,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:57,097 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:14:57,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:14:57,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:14:57,694 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:58,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:58,990 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:14:59,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:14:59,803 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:00,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:01,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:02,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:03,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:03,894 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:04,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:04,990 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:05,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:06,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:07,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:08,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:09,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:09,990 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:10,271 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:10,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:11,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:12,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:15:12,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:15:12,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:13,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:14,204 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:14,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:14,991 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:15,210 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:15,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:16,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:17,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:18,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:19,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:20,120 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:20,428 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:20,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:21,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:22,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:23,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:24,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:25,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:25,610 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:25,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:26,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:27,099 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:15:27,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:15:27,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:15:27,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:28,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:29,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:30,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:30,837 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:31,369 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:31,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:32,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:33,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:34,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:35,356 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:35,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:36,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:36,834 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:37,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:38,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:39,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:40,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:41,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:41,885 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:42,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:15:42,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:15:42,547 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:42,698 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:43,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:44,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:45,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:46,247 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:46,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:46,925 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:47,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:48,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:49,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:50,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:51,448 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:51,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:52,261 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:52,699 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:53,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:54,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:55,373 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:55,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:56,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:57,110 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:15:57,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:15:57,166 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:15:57,619 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:15:57,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:58,244 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:15:58,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:15:59,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:00,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:01,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:01,723 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:02,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:03,642 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:03,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:04,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:05,700 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:06,701 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:07,570 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:07,701 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:08,701 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:08,839 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:09,701 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:10,701 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:11,701 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:12,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:16:12,165 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:16:12,701 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:13,647 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:13,701 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:13,992 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:14,706 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:15,706 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:15,760 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:16,706 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:17,706 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:18,706 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:18,992 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:19,536 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:19,706 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:20,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:21,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:22,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:23,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:24,096 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:24,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:25,585 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:25,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:26,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:27,103 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:16:27,165 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:16:27,166 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:16:27,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:28,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:29,147 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:29,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:30,707 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:31,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:31,890 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:32,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:33,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:34,183 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:34,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:35,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:35,882 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:36,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:37,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:38,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:39,306 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:39,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:40,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:41,708 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:42,162 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:42,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:16:42,166 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:16:42,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:43,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:44,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:44,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:45,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:46,030 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:46,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:47,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:48,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:49,415 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:49,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:50,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:51,316 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:51,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:52,709 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:53,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:54,059 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:54,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:54,992 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:16:55,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:56,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:57,106 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:16:57,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:16:57,166 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:16:57,663 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:16:57,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:58,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:16:59,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:00,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:00,952 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:01,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:02,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:03,703 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:03,710 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:04,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:05,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:06,002 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:06,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:07,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:08,547 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:08,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:09,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:10,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:11,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:11,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:12,080 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:12,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:17:12,166 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:17:12,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:13,322 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:13,711 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:14,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:15,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:16,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:16,982 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:17,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:18,139 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:18,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:19,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:20,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:21,548 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:21,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:22,677 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:22,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:23,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:23,891 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:24,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:25,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:26,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:27,107 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:17:27,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:17:27,166 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:17:27,301 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:27,712 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:28,314 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:28,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:29,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:30,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:31,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:32,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:33,481 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:33,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:33,888 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:34,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:35,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:36,713 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:37,456 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:37,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:38,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:38,993 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:39,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:40,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:41,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:41,903 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:42,166 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:17:42,166 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:17:42,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:43,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:44,146 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:44,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:44,993 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:45,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:46,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:47,714 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:48,243 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:48,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:49,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:50,535 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:50,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:50,993 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:51,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:52,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:53,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:53,958 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:17:54,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:55,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:56,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:56,994 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:17:57,109 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:17:57,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:17:57,170 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:17:57,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:58,715 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:59,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:17:59,959 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:00,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:01,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:02,523 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:02,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:03,329 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:03,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:04,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:05,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:06,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:07,582 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:07,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:08,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:09,716 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:10,222 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:10,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:11,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:12,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:18:12,170 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:18:12,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:13,232 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:13,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:14,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:15,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:16,443 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:16,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:17,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:18,359 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:18,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:19,717 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:20,082 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:20,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:21,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:22,359 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:22,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:23,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:23,803 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:24,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:25,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:26,079 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:26,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:27,112 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:18:27,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:18:27,170 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:18:27,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:28,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:28,867 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:29,448 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:29,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:30,718 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:31,508 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:31,719 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:32,719 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:33,719 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:33,941 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:34,719 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:35,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:36,179 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:36,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:37,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:38,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:38,994 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:39,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:39,834 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:40,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:41,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:42,003 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:42,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:18:42,170 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:18:42,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:43,721 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:43,994 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:44,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:45,372 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:45,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:46,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:47,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:48,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:48,994 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:49,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:50,223 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:50,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:51,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:52,459 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:52,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:53,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:53,994 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:54,722 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:55,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:55,836 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:56,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:57,114 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:18:57,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:18:57,171 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:18:57,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:58,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:18:58,994 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:18:59,611 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:18:59,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:00,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:01,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:01,812 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:02,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:03,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:04,021 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:04,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:05,723 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:06,291 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:06,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:07,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:08,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:09,038 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:09,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:09,748 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:10,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:11,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:11,891 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:12,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:19:12,171 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:19:12,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:13,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:14,071 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:14,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:15,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:16,049 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:16,724 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:17,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:18,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:19,339 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:19,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:20,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:21,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:21,779 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:22,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:23,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:24,385 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:24,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:25,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:26,299 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:26,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:27,116 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:19:27,170 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:19:27,171 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:19:27,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:28,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:29,494 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:29,725 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:30,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:31,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:32,019 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:32,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:33,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:34,549 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:34,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:35,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:35,771 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:36,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:37,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:37,919 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:38,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:39,641 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:39,726 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:40,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:41,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:42,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:19:42,171 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:19:42,487 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:42,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:43,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:44,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:45,516 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:45,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:46,181 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:46,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:47,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:48,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:49,652 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:49,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:50,727 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:51,096 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:51,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:51,987 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:52,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:53,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:54,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:55,435 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:19:55,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:56,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:56,996 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:19:57,118 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:19:57,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:19:57,171 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:19:57,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:58,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:19:59,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:00,351 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:20:00,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:01,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:02,384 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:02,728 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:03,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:03,759 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:20:04,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:05,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:06,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:07,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:08,234 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:08,403 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:20:08,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:09,523 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:20:09,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:10,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:11,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:12,171 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:20:12,171 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:20:12,729 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:13,328 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:13,730 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:14,287 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:20:14,730 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:15,730 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:16,730 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:17,667 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:20:17,731 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:18,644 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:18,733 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:19,733 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:20,734 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:21,734 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:22,735 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:23,736 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:23,997 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:24,736 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:25,736 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:26,846 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:27,120 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:20:27,177 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:20:27,177 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:20:27,911 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:29,245 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:29,309 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:30,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:31,310 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:32,377 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:33,253 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:34,638 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:35,308 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:35,488 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:36,661 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:37,914 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:38,958 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:40,016 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:40,448 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:41,045 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:42,210 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:42,211 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:20:42,211 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:20:43,265 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:44,331 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:45,504 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:45,509 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:46,380 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:47,630 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:48,650 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:49,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:50,697 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:50,995 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:52,048 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:53,173 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:54,198 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:55,464 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:55,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:20:56,667 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:57,123 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:20:57,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:20:57,248 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:20:57,695 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:58,696 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:20:59,823 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:00,503 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:21:00,824 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:01,613 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:21:01,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:01,879 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:03,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:04,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:05,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:06,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:06,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:07,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:08,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:09,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:10,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:11,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:12,057 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:12,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:21:12,247 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:21:12,363 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:13,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:14,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:15,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:16,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:17,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:17,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:18,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:19,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:20,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:21,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:22,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:22,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:23,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:24,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:25,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:26,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:27,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:27,124 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:21:27,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:21:27,247 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:21:28,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:28,330 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:29,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:30,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:31,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:32,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:33,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:33,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:34,059 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:35,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:36,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:37,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:38,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:39,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:39,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:40,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:41,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:42,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:42,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:21:42,247 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:21:43,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:44,060 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:44,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:45,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:46,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:47,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:48,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:49,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:49,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:50,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:51,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:52,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:53,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:54,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:55,061 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:55,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:21:56,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:57,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:57,126 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:21:57,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:21:57,247 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:21:58,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:21:59,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:00,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:00,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:01,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:02,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:03,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:04,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:05,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:05,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:06,062 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:07,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:08,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:09,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:10,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:11,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:11,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:12,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:12,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:22:12,248 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:22:13,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:14,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:15,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:16,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:16,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:17,063 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:18,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:19,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:20,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:21,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:22,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:22,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:23,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:24,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:25,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:26,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:27,064 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:27,129 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:22:27,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:22:27,248 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:22:28,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:28,418 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:29,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:30,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:31,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:32,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:33,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:33,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:34,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:35,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:36,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:37,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:38,065 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:39,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:39,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:40,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:41,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:42,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:42,247 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:22:42,248 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:22:43,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:44,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:44,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:45,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:46,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:47,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:48,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:49,066 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:49,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:50,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:51,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:52,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:53,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:54,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:54,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:22:55,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:56,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:57,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:57,131 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:22:57,248 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:22:57,248 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:22:58,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:22:59,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:00,067 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:00,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:23:01,068 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:02,068 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:02,686 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:23:03,068 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:04,068 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:04,160 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:23:05,068 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:06,058 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:23:06,068 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:06,874 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:23:07,642 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:08,199 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:23:08,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:09,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:10,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:11,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:11,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:23:12,248 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: stop_status +2024-06-08 12:23:12,248 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: stop_status +2024-06-08 12:23:12,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:13,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:14,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:15,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:16,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:16,833 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:23:17,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:18,576 DEBUG SenderThread:989 [sender.py:send():379] send: config +2024-06-08 12:23:18,621 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: partial_history +2024-06-08 12:23:18,624 DEBUG SenderThread:989 [sender.py:send():379] send: telemetry +2024-06-08 12:23:18,624 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: summary_record +2024-06-08 12:23:18,625 INFO SenderThread:989 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 12:23:18,626 DEBUG SenderThread:989 [sender.py:send():379] send: history +2024-06-08 12:23:18,626 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: summary_record +2024-06-08 12:23:18,627 INFO SenderThread:989 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 12:23:18,829 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: log_artifact +2024-06-08 12:23:18,830 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: log_artifact +2024-06-08 12:23:18,843 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:18,979 INFO Thread-12 :989 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-summary.json +2024-06-08 12:23:18,979 INFO Thread-12 :989 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json +2024-06-08 12:23:18,979 INFO Thread-12 :989 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/media/table +2024-06-08 12:23:18,979 INFO Thread-12 :989 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/media/table/evaluation +2024-06-08 12:23:18,979 INFO Thread-12 :989 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/media +2024-06-08 12:23:18,984 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: partial_history +2024-06-08 12:23:19,083 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: log_artifact +2024-06-08 12:23:19,220 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: server_info +2024-06-08 12:23:19,418 INFO wandb-upload_0:989 [upload_job.py:push():130] Uploaded file /tmp/tmpjx21gu8gwandb/xmo2sh96-media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json +2024-06-08 12:23:19,650 INFO wandb-upload_1:989 [upload_job.py:push():88] Uploaded file /root/.local/share/wandb/artifacts/staging/tmpnpwci0t9 +2024-06-08 12:23:20,177 INFO SenderThread:989 [sender.py:send_request_log_artifact():1518] logged artifact run-9apxn9eo-evaluationeval_results - {'id': 'QXJ0aWZhY3Q6ODY0NzgyMzM5', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE4MzYzNjY5NQ==', 'latestArtifact': None}} +2024-06-08 12:23:20,177 DEBUG SenderThread:989 [sender.py:send():379] send: files +2024-06-08 12:23:20,177 INFO SenderThread:989 [sender.py:_save_file():1454] saving file media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json with policy now +2024-06-08 12:23:20,178 DEBUG SenderThread:989 [sender.py:send():379] send: history +2024-06-08 12:23:20,178 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: summary_record +2024-06-08 12:23:20,181 INFO SenderThread:989 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 12:23:20,181 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: log_artifact +2024-06-08 12:23:20,528 INFO wandb-upload_0:989 [upload_job.py:push():130] Uploaded file /tmp/tmpjx21gu8gwandb/ryzciq9v-media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json +2024-06-08 12:23:20,830 INFO wandb-upload_1:989 [upload_job.py:push():88] Uploaded file /tmp/tmpr7jxut4g/results.json +2024-06-08 12:23:20,982 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-summary.json +2024-06-08 12:23:21,322 INFO SenderThread:989 [sender.py:send_request_log_artifact():1518] logged artifact results - {'id': 'QXJ0aWZhY3Q6ODY0NzgyMzUw', 'state': 'PENDING', 'artifactSequence': {'id': 'QXJ0aWZhY3RDb2xsZWN0aW9uOjE4MjQwOTE0OQ==', 'latestArtifact': {'id': 'QXJ0aWZhY3Q6ODYyMTk3MTg2', 'versionIndex': 8}}} +2024-06-08 12:23:21,322 DEBUG SenderThread:989 [sender.py:send():379] send: telemetry +2024-06-08 12:23:21,322 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: server_info +2024-06-08 12:23:21,387 DEBUG SenderThread:989 [sender.py:send():379] send: exit +2024-06-08 12:23:21,387 INFO SenderThread:989 [sender.py:send_exit():586] handling exit code: 0 +2024-06-08 12:23:21,388 INFO SenderThread:989 [sender.py:send_exit():588] handling runtime: 4374 +2024-06-08 12:23:21,390 INFO SenderThread:989 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 12:23:21,390 INFO SenderThread:989 [sender.py:send_exit():594] send defer +2024-06-08 12:23:21,390 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:21,390 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 0 +2024-06-08 12:23:21,391 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:21,391 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 0 +2024-06-08 12:23:21,391 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 1 +2024-06-08 12:23:21,391 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:21,391 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 1 +2024-06-08 12:23:21,391 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:21,391 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 1 +2024-06-08 12:23:21,391 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 2 +2024-06-08 12:23:21,391 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:21,391 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 2 +2024-06-08 12:23:21,391 INFO HandlerThread:989 [system_monitor.py:finish():203] Stopping system monitor +2024-06-08 12:23:21,391 DEBUG SystemMonitor:989 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-06-08 12:23:21,391 DEBUG SystemMonitor:989 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-06-08 12:23:21,394 INFO HandlerThread:989 [interfaces.py:finish():200] Joined cpu monitor +2024-06-08 12:23:21,394 INFO HandlerThread:989 [interfaces.py:finish():200] Joined disk monitor +2024-06-08 12:23:21,394 INFO HandlerThread:989 [interfaces.py:finish():200] Joined memory monitor +2024-06-08 12:23:21,394 INFO HandlerThread:989 [interfaces.py:finish():200] Joined network monitor +2024-06-08 12:23:21,395 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:21,395 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 2 +2024-06-08 12:23:21,395 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 3 +2024-06-08 12:23:21,395 DEBUG SenderThread:989 [sender.py:send():379] send: stats +2024-06-08 12:23:21,395 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:21,396 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 3 +2024-06-08 12:23:21,396 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:21,396 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 3 +2024-06-08 12:23:21,396 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 4 +2024-06-08 12:23:21,396 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:21,396 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 4 +2024-06-08 12:23:21,396 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:21,396 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 4 +2024-06-08 12:23:21,396 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 5 +2024-06-08 12:23:21,396 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:21,396 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 5 +2024-06-08 12:23:21,397 DEBUG SenderThread:989 [sender.py:send():379] send: summary +2024-06-08 12:23:21,398 INFO SenderThread:989 [sender.py:_save_file():1454] saving file wandb-summary.json with policy end +2024-06-08 12:23:21,398 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:21,398 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 5 +2024-06-08 12:23:21,398 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 6 +2024-06-08 12:23:21,398 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:21,398 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 6 +2024-06-08 12:23:21,398 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:21,398 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 6 +2024-06-08 12:23:21,402 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: status_report +2024-06-08 12:23:21,529 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 7 +2024-06-08 12:23:21,530 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:21,530 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 7 +2024-06-08 12:23:21,530 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:21,530 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 7 +2024-06-08 12:23:21,984 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/config.yaml +2024-06-08 12:23:21,985 INFO Thread-12 :989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-summary.json +2024-06-08 12:23:22,388 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: poll_exit +2024-06-08 12:23:22,849 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 8 +2024-06-08 12:23:22,850 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: poll_exit +2024-06-08 12:23:22,850 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:22,850 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 8 +2024-06-08 12:23:22,850 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:22,850 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 8 +2024-06-08 12:23:22,850 INFO SenderThread:989 [job_builder.py:build():440] Attempting to build job artifact +2024-06-08 12:23:22,853 INFO SenderThread:989 [job_builder.py:_get_source_type():580] no source found +2024-06-08 12:23:22,853 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 9 +2024-06-08 12:23:22,853 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:22,854 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 9 +2024-06-08 12:23:22,854 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:22,854 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 9 +2024-06-08 12:23:22,854 INFO SenderThread:989 [dir_watcher.py:finish():358] shutting down directory watcher +2024-06-08 12:23:22,986 INFO SenderThread:989 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:23:22,987 INFO SenderThread:989 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files +2024-06-08 12:23:22,987 INFO SenderThread:989 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/config.yaml config.yaml +2024-06-08 12:23:22,987 INFO SenderThread:989 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log output.log +2024-06-08 12:23:22,989 INFO SenderThread:989 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-summary.json wandb-summary.json +2024-06-08 12:23:22,991 INFO SenderThread:989 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-metadata.json wandb-metadata.json +2024-06-08 12:23:22,991 INFO SenderThread:989 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/requirements.txt requirements.txt +2024-06-08 12:23:22,992 INFO SenderThread:989 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json media/table/evaluation/eval_results_1_6529e3311149275b8699.table.json +2024-06-08 12:23:22,992 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 10 +2024-06-08 12:23:22,992 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:22,992 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 10 +2024-06-08 12:23:22,992 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:22,992 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 10 +2024-06-08 12:23:22,992 INFO SenderThread:989 [file_pusher.py:finish():169] shutting down file pusher +2024-06-08 12:23:23,258 INFO wandb-upload_0:989 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/config.yaml +2024-06-08 12:23:23,346 INFO wandb-upload_1:989 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/output.log +2024-06-08 12:23:23,387 INFO wandb-upload_3:989 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/requirements.txt +2024-06-08 12:23:23,389 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: poll_exit +2024-06-08 12:23:23,389 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: poll_exit +2024-06-08 12:23:23,396 INFO wandb-upload_2:989 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/files/wandb-summary.json +2024-06-08 12:23:23,597 INFO Thread-11 (_thread_body):989 [sender.py:transition_state():614] send defer: 11 +2024-06-08 12:23:23,597 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:23,597 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 11 +2024-06-08 12:23:23,597 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:23,597 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 11 +2024-06-08 12:23:23,597 INFO SenderThread:989 [file_pusher.py:join():175] waiting for file pusher +2024-06-08 12:23:23,598 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 12 +2024-06-08 12:23:23,598 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:23,598 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 12 +2024-06-08 12:23:23,598 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:23,598 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 12 +2024-06-08 12:23:23,598 INFO SenderThread:989 [file_stream.py:finish():601] file stream finish called +2024-06-08 12:23:23,828 INFO SenderThread:989 [file_stream.py:finish():605] file stream finish is done +2024-06-08 12:23:23,828 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 13 +2024-06-08 12:23:23,828 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:23,828 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 13 +2024-06-08 12:23:23,829 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:23,829 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 13 +2024-06-08 12:23:23,829 INFO SenderThread:989 [sender.py:transition_state():614] send defer: 14 +2024-06-08 12:23:23,829 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: defer +2024-06-08 12:23:23,829 INFO HandlerThread:989 [handler.py:handle_request_defer():184] handle defer: 14 +2024-06-08 12:23:23,829 DEBUG SenderThread:989 [sender.py:send():379] send: final +2024-06-08 12:23:23,829 DEBUG SenderThread:989 [sender.py:send():379] send: footer +2024-06-08 12:23:23,829 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: defer +2024-06-08 12:23:23,829 INFO SenderThread:989 [sender.py:send_request_defer():610] handle sender defer: 14 +2024-06-08 12:23:23,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: poll_exit +2024-06-08 12:23:23,830 DEBUG SenderThread:989 [sender.py:send_request():406] send_request: poll_exit +2024-06-08 12:23:23,830 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: internal_messages +2024-06-08 12:23:23,831 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: get_summary +2024-06-08 12:23:23,832 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: sampled_history +2024-06-08 12:23:23,833 DEBUG HandlerThread:989 [handler.py:handle_request():158] handle_request: shutdown +2024-06-08 12:23:23,833 INFO HandlerThread:989 [handler.py:finish():882] shutting down handler +2024-06-08 12:23:24,830 INFO WriterThread:989 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/run-9apxn9eo.wandb +2024-06-08 12:23:24,830 INFO SenderThread:989 [sender.py:finish():1608] shutting down sender +2024-06-08 12:23:24,830 INFO SenderThread:989 [file_pusher.py:finish():169] shutting down file pusher +2024-06-08 12:23:24,830 INFO SenderThread:989 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/logs/debug.log b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..8b91a77576488d2efcb56cf835b38b805e94891c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/logs/debug.log @@ -0,0 +1,36 @@ +2024-06-08 11:10:26,457 INFO MainThread:833 [wandb_setup.py:_flush():76] Current SDK version is 0.17.1 +2024-06-08 11:10:26,457 INFO MainThread:833 [wandb_setup.py:_flush():76] Configure stats pid to 833 +2024-06-08 11:10:26,457 INFO MainThread:833 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-06-08 11:10:26,457 INFO MainThread:833 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-06-08 11:10:26,458 WARNING MainThread:833 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/logs/debug.log +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240608_111026-9apxn9eo/logs/debug-internal.log +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_init.py:init():560] calling init triggers +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_init.py:init():610] starting backend +2024-06-08 11:10:26,458 INFO MainThread:833 [wandb_init.py:init():614] setting up manager +2024-06-08 11:10:26,461 INFO MainThread:833 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-06-08 11:10:26,462 INFO MainThread:833 [wandb_init.py:init():622] backend started and connected +2024-06-08 11:10:26,466 INFO MainThread:833 [wandb_init.py:init():711] updated telemetry +2024-06-08 11:10:26,475 INFO MainThread:833 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-06-08 11:10:26,724 INFO MainThread:833 [wandb_run.py:_on_init():2402] communicating current version +2024-06-08 11:10:26,754 INFO MainThread:833 [wandb_run.py:_on_init():2411] got version response +2024-06-08 11:10:26,754 INFO MainThread:833 [wandb_init.py:init():795] starting run threads in backend +2024-06-08 11:10:27,095 INFO MainThread:833 [wandb_run.py:_console_start():2380] atexit reg +2024-06-08 11:10:27,095 INFO MainThread:833 [wandb_run.py:_redirect():2235] redirect: wrap_raw +2024-06-08 11:10:27,095 INFO MainThread:833 [wandb_run.py:_redirect():2300] Wrapping output streams. +2024-06-08 11:10:27,095 INFO MainThread:833 [wandb_run.py:_redirect():2325] Redirects installed. +2024-06-08 11:10:27,099 INFO MainThread:833 [wandb_init.py:init():838] run started, returning control to user process +2024-06-08 12:23:18,575 INFO MainThread:833 [wandb_run.py:_config_callback():1382] config_cb None None {'task_configs': {'arc_easy': {'task': 'arc_easy', 'group': ['ai2_arc'], 'dataset_path': 'allenai/ai2_arc', 'dataset_name': 'ARC-Easy', 'training_split': 'train', 'validation_split': 'validation', 'test_split': 'test', 'doc_to_text': 'Question: {{question}}\nAnswer:', 'doc_to_target': '{{choices.label.index(answerKey)}}', 'doc_to_choice': '{{choices.text}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}, {'metric': 'acc_norm', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'Question: {{question}}\nAnswer:', 'metadata': {'version': 1.0}}, 'boolq': {'task': 'boolq', 'group': ['super-glue-lm-eval-v1'], 'dataset_path': 'super_glue', 'dataset_name': 'boolq', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': '{{passage}}\nQuestion: {{question}}?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['no', 'yes'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'passage', 'metadata': {'version': 2.0}}, 'copa': {'task': 'copa', 'group': ['super-glue-lm-eval-v1'], 'dataset_path': 'super_glue', 'dataset_name': 'copa', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'def doc_to_text(doc):\n # Drop the period\n connector = {\n "cause": "because",\n "effect": "therefore",\n }[doc["question"]]\n return doc["premise"].strip()[:-1] + f" {connector}"\n', 'doc_to_target': 'def doc_to_target(doc):\n correct_choice = doc["choice1"] if doc["label"] == 0 else doc["choice2"]\n # Connect the sentences\n return " " + convert_choice(correct_choice)\n', 'doc_to_choice': 'def doc_to_choice(doc):\n return [" " + convert_choice(doc["choice1"]), " " + convert_choice(doc["choice2"])]\n', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'indic_arc_challenge_hi': {'task': 'indic_arc_challenge_hi', 'group': 'Cognitive-Lab/Indic-ARC-Challenge', 'dataset_path': 'Cognitive-Lab/Indic-ARC-Challenge', 'dataset_name': 'hi', 'test_split': 'test', 'doc_to_text': 'Question: {{translated_question}}\nAnswer:', 'doc_to_target': '{{translated_choices.label.index(answerKey)}}', 'doc_to_choice': '{{translated_choices.text}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'Question: {{translated_question}}\nAnswer:', 'metadata': {'version': 1.0}}, 'indic_arc_easy_hi': {'task': 'indic_arc_easy_hi', 'group': 'Cognitive-Lab/Indic-ARC-Easy', 'dataset_path': 'Cognitive-Lab/Indic-ARC-Easy', 'dataset_name': 'hi', 'test_split': 'test', 'doc_to_text': 'Question: {{translated_question}}\nAnswer:', 'doc_to_target': '{{translated_choices.label.index(answerKey)}}', 'doc_to_choice': '{{translated_choices.text}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'Question: {{translated_question}}\nAnswer:', 'metadata': {'version': 1.0}}, 'indic_boolq_hi': {'task': 'indic_boolq_hi', 'group': 'Cognitive-Lab/Indic-BoolQ', 'dataset_path': 'Cognitive-Lab/Indic-BoolQ', 'dataset_name': 'hi', 'validation_split': 'validation', 'doc_to_text': 'Passage: {translated_passage}\nQuestion: {translated_question.strip()}\nAnswer:', 'doc_to_target': 'answer', 'doc_to_choice': ['true', 'false'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'mrpc': {'task': 'mrpc', 'group': 'glue', 'dataset_path': 'glue', 'dataset_name': 'mrpc', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'Sentence 1: {{sentence1}}\nSentence 2: {{sentence2}}\nQuestion: Do both sentences mean the same thing?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['no', 'yes'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}, {'metric': 'f1'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'piqa': {'task': 'piqa', 'dataset_path': 'piqa', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'Question: {{goal}}\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': '{{[sol1, sol2]}}', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}, {'metric': 'acc_norm', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'goal', 'metadata': {'version': 1.0}}, 'sst2': {'task': 'sst2', 'group': 'glue', 'dataset_path': 'glue', 'dataset_name': 'sst2', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': '{{sentence}}\nQuestion: Is this sentence positive or negative?\nAnswer:', 'doc_to_target': 'label', 'doc_to_choice': ['negative', 'positive'], 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc'}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': False, 'metadata': {'version': 1.0}}, 'winogrande': {'task': 'winogrande', 'dataset_path': 'winogrande', 'dataset_name': 'winogrande_xl', 'training_split': 'train', 'validation_split': 'validation', 'doc_to_text': 'def doc_to_text(doc):\n answer_to_num = {"1": 0, "2": 1}\n return answer_to_num[doc["answer"]]\n', 'doc_to_target': 'def doc_to_target(doc):\n idx = doc["sentence"].index("_") + 1\n return doc["sentence"][idx:].strip()\n', 'doc_to_choice': 'def doc_to_choice(doc):\n idx = doc["sentence"].index("_")\n options = [doc["option1"], doc["option2"]]\n return [doc["sentence"][:idx] + opt for opt in options]\n', 'description': '', 'target_delimiter': ' ', 'fewshot_delimiter': '\n\n', 'num_fewshot': 0, 'metric_list': [{'metric': 'acc', 'aggregation': 'mean', 'higher_is_better': True}], 'output_type': 'multiple_choice', 'repeats': 1, 'should_decontaminate': True, 'doc_to_decontamination_query': 'sentence', 'metadata': {'version': 1.0}}}, 'cli_configs': {'model': 'hf', 'model_args': 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn-updated/llamav2-3b/hf/global_step100000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer', 'batch_size': 'auto', 'batch_sizes': [64], 'device': None, 'use_cache': None, 'limit': None, 'bootstrap_iters': 100000, 'gen_kwargs': None}} +2024-06-08 12:23:19,219 INFO MainThread:833 [wandb_run.py:_finish():2109] finishing run smlgenai/engl-hi-eval/9apxn9eo +2024-06-08 12:23:19,219 INFO MainThread:833 [wandb_run.py:_atexit_cleanup():2349] got exitcode: 0 +2024-06-08 12:23:19,220 INFO MainThread:833 [wandb_run.py:_restore():2332] restore +2024-06-08 12:23:19,220 INFO MainThread:833 [wandb_run.py:_restore():2338] restore done +2024-06-08 12:23:24,834 INFO MainThread:833 [wandb_run.py:_footer_history_summary_info():4008] rendering history +2024-06-08 12:23:24,835 INFO MainThread:833 [wandb_run.py:_footer_history_summary_info():4040] rendering summary +2024-06-08 12:23:24,841 INFO MainThread:833 [wandb_run.py:_footer_sync_info():3967] logging synced files diff --git a/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/configuration_align.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/configuration_align.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..5728f6ba4b6d7235f16c3e09c638996a779a693e Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/configuration_align.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/convert_align_tf_to_hf.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/convert_align_tf_to_hf.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..bd7a479e3dc03c7930cba7ea37eb39ae6da2c78d Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/convert_align_tf_to_hf.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/modeling_align.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/modeling_align.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..cf40d3356707d35a9bd2bcd6941b7e4ee66a3dc4 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/modeling_align.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/processing_align.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/processing_align.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..f6d84b88875f2f257b381acee8f08900c9e1f9ca Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/align/__pycache__/processing_align.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/align/configuration_align.py b/venv/lib/python3.10/site-packages/transformers/models/align/configuration_align.py new file mode 100644 index 0000000000000000000000000000000000000000..a4b3149d971a15145c6000c31fda6f020f06e0af --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/align/configuration_align.py @@ -0,0 +1,383 @@ +# coding=utf-8 +# Copyright 2023 The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" ALIGN model configuration""" + +import os +from typing import TYPE_CHECKING, List, Union + + +if TYPE_CHECKING: + pass + +from ...configuration_utils import PretrainedConfig +from ...utils import logging + + +logger = logging.get_logger(__name__) + + +from ..deprecated._archive_maps import ALIGN_PRETRAINED_CONFIG_ARCHIVE_MAP # noqa: F401, E402 + + +class AlignTextConfig(PretrainedConfig): + r""" + This is the configuration class to store the configuration of a [`AlignTextModel`]. It is used to instantiate a + ALIGN text encoder according to the specified arguments, defining the model architecture. Instantiating a + configuration with the defaults will yield a similar configuration to that of the text encoder of the ALIGN + [kakaobrain/align-base](https://huggingface.co/kakaobrain/align-base) architecture. The default values here are + copied from BERT. + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + Args: + vocab_size (`int`, *optional*, defaults to 30522): + Vocabulary size of the Align Text model. Defines the number of different tokens that can be represented by + the `inputs_ids` passed when calling [`AlignTextModel`]. + hidden_size (`int`, *optional*, defaults to 768): + Dimensionality of the encoder layers and the pooler layer. + num_hidden_layers (`int`, *optional*, defaults to 12): + Number of hidden layers in the Transformer encoder. + num_attention_heads (`int`, *optional*, defaults to 12): + Number of attention heads for each attention layer in the Transformer encoder. + intermediate_size (`int`, *optional*, defaults to 3072): + Dimensionality of the "intermediate" (often named feed-forward) layer in the Transformer encoder. + hidden_act (`str` or `Callable`, *optional*, defaults to `"gelu"`): + The non-linear activation function (function or string) in the encoder and pooler. If string, `"gelu"`, + `"relu"`, `"silu"` and `"gelu_new"` are supported. + hidden_dropout_prob (`float`, *optional*, defaults to 0.1): + The dropout probability for all fully connected layers in the embeddings, encoder, and pooler. + attention_probs_dropout_prob (`float`, *optional*, defaults to 0.1): + The dropout ratio for the attention probabilities. + max_position_embeddings (`int`, *optional*, defaults to 512): + The maximum sequence length that this model might ever be used with. Typically set this to something large + just in case (e.g., 512 or 1024 or 2048). + type_vocab_size (`int`, *optional*, defaults to 2): + The vocabulary size of the `token_type_ids` passed when calling [`AlignTextModel`]. + initializer_range (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + layer_norm_eps (`float`, *optional*, defaults to 1e-12): + The epsilon used by the layer normalization layers. + pad_token_id (`int`, *optional*, defaults to 0): + Padding token id. + position_embedding_type (`str`, *optional*, defaults to `"absolute"`): + Type of position embedding. Choose one of `"absolute"`, `"relative_key"`, `"relative_key_query"`. For + positional embeddings use `"absolute"`. For more information on `"relative_key"`, please refer to + [Self-Attention with Relative Position Representations (Shaw et al.)](https://arxiv.org/abs/1803.02155). + For more information on `"relative_key_query"`, please refer to *Method 4* in [Improve Transformer Models + with Better Relative Position Embeddings (Huang et al.)](https://arxiv.org/abs/2009.13658). + use_cache (`bool`, *optional*, defaults to `True`): + Whether or not the model should return the last key/values attentions (not used by all models). Only + relevant if `config.is_decoder=True`. + + Example: + + ```python + >>> from transformers import AlignTextConfig, AlignTextModel + + >>> # Initializing a AlignTextConfig with kakaobrain/align-base style configuration + >>> configuration = AlignTextConfig() + + >>> # Initializing a AlignTextModel (with random weights) from the kakaobrain/align-base style configuration + >>> model = AlignTextModel(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + ```""" + + model_type = "align_text_model" + + def __init__( + self, + vocab_size=30522, + hidden_size=768, + num_hidden_layers=12, + num_attention_heads=12, + intermediate_size=3072, + hidden_act="gelu", + hidden_dropout_prob=0.1, + attention_probs_dropout_prob=0.1, + max_position_embeddings=512, + type_vocab_size=2, + initializer_range=0.02, + layer_norm_eps=1e-12, + pad_token_id=0, + position_embedding_type="absolute", + use_cache=True, + **kwargs, + ): + super().__init__(**kwargs) + + self.vocab_size = vocab_size + self.hidden_size = hidden_size + self.num_hidden_layers = num_hidden_layers + self.num_attention_heads = num_attention_heads + self.hidden_act = hidden_act + self.intermediate_size = intermediate_size + self.hidden_dropout_prob = hidden_dropout_prob + self.attention_probs_dropout_prob = attention_probs_dropout_prob + self.max_position_embeddings = max_position_embeddings + self.type_vocab_size = type_vocab_size + self.initializer_range = initializer_range + self.layer_norm_eps = layer_norm_eps + self.position_embedding_type = position_embedding_type + self.use_cache = use_cache + self.pad_token_id = pad_token_id + + @classmethod + def from_pretrained(cls, pretrained_model_name_or_path: Union[str, os.PathLike], **kwargs) -> "PretrainedConfig": + cls._set_token_in_kwargs(kwargs) + + config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs) + + # get the text config dict if we are loading from AlignConfig + if config_dict.get("model_type") == "align": + config_dict = config_dict["text_config"] + + if "model_type" in config_dict and hasattr(cls, "model_type") and config_dict["model_type"] != cls.model_type: + logger.warning( + f"You are using a model of type {config_dict['model_type']} to instantiate a model of type " + f"{cls.model_type}. This is not supported for all configurations of models and can yield errors." + ) + + return cls.from_dict(config_dict, **kwargs) + + +class AlignVisionConfig(PretrainedConfig): + r""" + This is the configuration class to store the configuration of a [`AlignVisionModel`]. It is used to instantiate a + ALIGN vision encoder according to the specified arguments, defining the model architecture. Instantiating a + configuration with the defaults will yield a similar configuration to that of the vision encoder of the ALIGN + [kakaobrain/align-base](https://huggingface.co/kakaobrain/align-base) architecture. The default values are copied + from EfficientNet (efficientnet-b7) + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + Args: + num_channels (`int`, *optional*, defaults to 3): + The number of input channels. + image_size (`int`, *optional*, defaults to 600): + The input image size. + width_coefficient (`float`, *optional*, defaults to 2.0): + Scaling coefficient for network width at each stage. + depth_coefficient (`float`, *optional*, defaults to 3.1): + Scaling coefficient for network depth at each stage. + depth_divisor `int`, *optional*, defaults to 8): + A unit of network width. + kernel_sizes (`List[int]`, *optional*, defaults to `[3, 3, 5, 3, 5, 5, 3]`): + List of kernel sizes to be used in each block. + in_channels (`List[int]`, *optional*, defaults to `[32, 16, 24, 40, 80, 112, 192]`): + List of input channel sizes to be used in each block for convolutional layers. + out_channels (`List[int]`, *optional*, defaults to `[16, 24, 40, 80, 112, 192, 320]`): + List of output channel sizes to be used in each block for convolutional layers. + depthwise_padding (`List[int]`, *optional*, defaults to `[]`): + List of block indices with square padding. + strides (`List[int]`, *optional*, defaults to `[1, 2, 2, 2, 1, 2, 1]`): + List of stride sizes to be used in each block for convolutional layers. + num_block_repeats (`List[int]`, *optional*, defaults to `[1, 2, 2, 3, 3, 4, 1]`): + List of the number of times each block is to repeated. + expand_ratios (`List[int]`, *optional*, defaults to `[1, 6, 6, 6, 6, 6, 6]`): + List of scaling coefficient of each block. + squeeze_expansion_ratio (`float`, *optional*, defaults to 0.25): + Squeeze expansion ratio. + hidden_act (`str` or `function`, *optional*, defaults to `"silu"`): + The non-linear activation function (function or string) in each block. If string, `"gelu"`, `"relu"`, + `"selu", `"gelu_new"`, `"silu"` and `"mish"` are supported. + hiddem_dim (`int`, *optional*, defaults to 1280): + The hidden dimension of the layer before the classification head. + pooling_type (`str` or `function`, *optional*, defaults to `"mean"`): + Type of final pooling to be applied before the dense classification head. Available options are [`"mean"`, + `"max"`] + initializer_range (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + batch_norm_eps (`float`, *optional*, defaults to 1e-3): + The epsilon used by the batch normalization layers. + batch_norm_momentum (`float`, *optional*, defaults to 0.99): + The momentum used by the batch normalization layers. + drop_connect_rate (`float`, *optional*, defaults to 0.2): + The drop rate for skip connections. + + Example: + + ```python + >>> from transformers import AlignVisionConfig, AlignVisionModel + + >>> # Initializing a AlignVisionConfig with kakaobrain/align-base style configuration + >>> configuration = AlignVisionConfig() + + >>> # Initializing a AlignVisionModel (with random weights) from the kakaobrain/align-base style configuration + >>> model = AlignVisionModel(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + ```""" + + model_type = "align_vision_model" + + def __init__( + self, + num_channels: int = 3, + image_size: int = 600, + width_coefficient: float = 2.0, + depth_coefficient: float = 3.1, + depth_divisor: int = 8, + kernel_sizes: List[int] = [3, 3, 5, 3, 5, 5, 3], + in_channels: List[int] = [32, 16, 24, 40, 80, 112, 192], + out_channels: List[int] = [16, 24, 40, 80, 112, 192, 320], + depthwise_padding: List[int] = [], + strides: List[int] = [1, 2, 2, 2, 1, 2, 1], + num_block_repeats: List[int] = [1, 2, 2, 3, 3, 4, 1], + expand_ratios: List[int] = [1, 6, 6, 6, 6, 6, 6], + squeeze_expansion_ratio: float = 0.25, + hidden_act: str = "swish", + hidden_dim: int = 2560, + pooling_type: str = "mean", + initializer_range: float = 0.02, + batch_norm_eps: float = 0.001, + batch_norm_momentum: float = 0.99, + drop_connect_rate: float = 0.2, + **kwargs, + ): + super().__init__(**kwargs) + + self.num_channels = num_channels + self.image_size = image_size + self.width_coefficient = width_coefficient + self.depth_coefficient = depth_coefficient + self.depth_divisor = depth_divisor + self.kernel_sizes = kernel_sizes + self.in_channels = in_channels + self.out_channels = out_channels + self.depthwise_padding = depthwise_padding + self.strides = strides + self.num_block_repeats = num_block_repeats + self.expand_ratios = expand_ratios + self.squeeze_expansion_ratio = squeeze_expansion_ratio + self.hidden_act = hidden_act + self.hidden_dim = hidden_dim + self.pooling_type = pooling_type + self.initializer_range = initializer_range + self.batch_norm_eps = batch_norm_eps + self.batch_norm_momentum = batch_norm_momentum + self.drop_connect_rate = drop_connect_rate + self.num_hidden_layers = sum(num_block_repeats) * 4 + + @classmethod + def from_pretrained(cls, pretrained_model_name_or_path: Union[str, os.PathLike], **kwargs) -> "PretrainedConfig": + cls._set_token_in_kwargs(kwargs) + + config_dict, kwargs = cls.get_config_dict(pretrained_model_name_or_path, **kwargs) + + # get the vision config dict if we are loading from AlignConfig + if config_dict.get("model_type") == "align": + config_dict = config_dict["vision_config"] + + if "model_type" in config_dict and hasattr(cls, "model_type") and config_dict["model_type"] != cls.model_type: + logger.warning( + f"You are using a model of type {config_dict['model_type']} to instantiate a model of type " + f"{cls.model_type}. This is not supported for all configurations of models and can yield errors." + ) + + return cls.from_dict(config_dict, **kwargs) + + +class AlignConfig(PretrainedConfig): + r""" + [`AlignConfig`] is the configuration class to store the configuration of a [`AlignModel`]. It is used to + instantiate a ALIGN model according to the specified arguments, defining the text model and vision model configs. + Instantiating a configuration with the defaults will yield a similar configuration to that of the ALIGN + [kakaobrain/align-base](https://huggingface.co/kakaobrain/align-base) architecture. + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + Args: + text_config (`dict`, *optional*): + Dictionary of configuration options used to initialize [`AlignTextConfig`]. + vision_config (`dict`, *optional*): + Dictionary of configuration options used to initialize [`AlignVisionConfig`]. + projection_dim (`int`, *optional*, defaults to 640): + Dimentionality of text and vision projection layers. + temperature_init_value (`float`, *optional*, defaults to 1.0): + The inital value of the *temperature* paramter. Default is used as per the original ALIGN implementation. + initializer_range (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + kwargs (*optional*): + Dictionary of keyword arguments. + + Example: + + ```python + >>> from transformers import AlignConfig, AlignModel + + >>> # Initializing a AlignConfig with kakaobrain/align-base style configuration + >>> configuration = AlignConfig() + + >>> # Initializing a AlignModel (with random weights) from the kakaobrain/align-base style configuration + >>> model = AlignModel(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + + >>> # We can also initialize a AlignConfig from a AlignTextConfig and a AlignVisionConfig + >>> from transformers import AlignTextConfig, AlignVisionConfig + + >>> # Initializing ALIGN Text and Vision configurations + >>> config_text = AlignTextConfig() + >>> config_vision = AlignVisionConfig() + + >>> config = AlignConfig.from_text_vision_configs(config_text, config_vision) + ```""" + + model_type = "align" + + def __init__( + self, + text_config=None, + vision_config=None, + projection_dim=640, + temperature_init_value=1.0, + initializer_range=0.02, + **kwargs, + ): + super().__init__(**kwargs) + + if text_config is None: + text_config = {} + logger.info("text_config is None. Initializing the AlignTextConfig with default values.") + + if vision_config is None: + vision_config = {} + logger.info("vision_config is None. Initializing the AlignVisionConfig with default values.") + + self.text_config = AlignTextConfig(**text_config) + self.vision_config = AlignVisionConfig(**vision_config) + + self.projection_dim = projection_dim + self.temperature_init_value = temperature_init_value + self.initializer_range = initializer_range + + @classmethod + def from_text_vision_configs(cls, text_config: AlignTextConfig, vision_config: AlignVisionConfig, **kwargs): + r""" + Instantiate a [`AlignConfig`] (or a derived class) from align text model configuration and align vision model + configuration. + + Returns: + [`AlignConfig`]: An instance of a configuration object + """ + + return cls(text_config=text_config.to_dict(), vision_config=vision_config.to_dict(), **kwargs) diff --git a/venv/lib/python3.10/site-packages/transformers/models/align/modeling_align.py b/venv/lib/python3.10/site-packages/transformers/models/align/modeling_align.py new file mode 100644 index 0000000000000000000000000000000000000000..3dce9d383da1516c481db264591bf880e7cb09c8 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/align/modeling_align.py @@ -0,0 +1,1633 @@ +# coding=utf-8 +# Copyright 2023 The Google Research Team Authors and The HuggingFace Team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" PyTorch ALIGN model.""" + +import math +from dataclasses import dataclass +from typing import Any, Optional, Tuple, Union + +import torch +import torch.utils.checkpoint +from torch import nn + +from ...activations import ACT2FN +from ...modeling_outputs import ( + BaseModelOutputWithNoAttention, + BaseModelOutputWithPastAndCrossAttentions, + BaseModelOutputWithPoolingAndCrossAttentions, + BaseModelOutputWithPoolingAndNoAttention, +) +from ...modeling_utils import PreTrainedModel +from ...pytorch_utils import apply_chunking_to_forward, find_pruneable_heads_and_indices, prune_linear_layer +from ...utils import ( + ModelOutput, + add_start_docstrings, + add_start_docstrings_to_model_forward, + logging, + replace_return_docstrings, +) +from .configuration_align import AlignConfig, AlignTextConfig, AlignVisionConfig + + +logger = logging.get_logger(__name__) + +_CHECKPOINT_FOR_DOC = "kakaobrain/align-base" +_CONFIG_FOR_DOC = "AlignConfig" + + +from ..deprecated._archive_maps import ALIGN_PRETRAINED_MODEL_ARCHIVE_LIST # noqa: F401, E402 + + +ALIGN_START_DOCSTRING = r""" + This model inherits from [`PreTrainedModel`]. Check the superclass documentation for the generic methods the + library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads + etc.) + + This model is also a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. + Use it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage + and behavior. + + Parameters: + config ([`AlignConfig`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. +""" + +ALIGN_TEXT_INPUTS_DOCSTRING = r""" + Args: + input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): + Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide + it. + + Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and + [`PreTrainedTokenizer.__call__`] for details. + + [What are input IDs?](../glossary#input-ids) + attention_mask (`torch.Tensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on padding token indices. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + [What are attention masks?](../glossary#attention-mask) + position_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): + Indices of positions of each input sequence tokens in the position embeddings. Selected in the range `[0, + config.max_position_embeddings - 1]`. + + [What are position IDs?](../glossary#position-ids) + token_type_ids (`torch.LongTensor` of shape `({0})`, *optional*): + Segment token indices to indicate first and second portions of the inputs. Indices are selected in `[0, + 1]`: + + - 0 corresponds to a *sentence A* token, + - 1 corresponds to a *sentence B* token. + + [What are token type IDs?](../glossary#token-type-ids) + head_mask (`torch.FloatTensor` of shape `(num_heads,)` or `(num_layers, num_heads)`, *optional*): + Mask to nullify selected heads of the self-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + inputs_embeds (`torch.FloatTensor` of shape `({0}, hidden_size)`, *optional*): + Optionally, instead of passing `input_ids` you can choose to directly pass an embedded representation. This + is useful if you want more control over how to convert `input_ids` indices into associated vectors than the + model's internal embedding lookup matrix. + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + +ALIGN_VISION_INPUTS_DOCSTRING = r""" + Args: + pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): + Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using + [`AutoImageProcessor`]. See [`EfficientNetImageProcessor.__call__`] for details. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + +ALIGN_INPUTS_DOCSTRING = r""" + Args: + input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`): + Indices of input sequence tokens in the vocabulary. Padding will be ignored by default should you provide + it. + + Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and + [`PreTrainedTokenizer.__call__`] for details. + + [What are input IDs?](../glossary#input-ids) + attention_mask (`torch.Tensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on padding token indices. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + [What are attention masks?](../glossary#attention-mask) + position_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): + Indices of positions of each input sequence tokens in the position embeddings. Selected in the range `[0, + config.max_position_embeddings - 1]`. + + [What are position IDs?](../glossary#position-ids) + token_type_ids (`torch.LongTensor` of shape `({0})`, *optional*): + Segment token indices to indicate first and second portions of the inputs. Indices are selected in `[0, + 1]`: + + - 0 corresponds to a *sentence A* token, + - 1 corresponds to a *sentence B* token. + + [What are token type IDs?](../glossary#token-type-ids) + head_mask (`torch.FloatTensor` of shape `(num_heads,)` or `(num_layers, num_heads)`, *optional*): + Mask to nullify selected heads of the self-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + inputs_embeds (`torch.FloatTensor` of shape `({0}, hidden_size)`, *optional*): + Optionally, instead of passing `input_ids` you can choose to directly pass an embedded representation. This + is useful if you want more control over how to convert `input_ids` indices into associated vectors than the + model's internal embedding lookup matrix. + pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): + Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using + [`AutoImageProcessor`]. See [`EfficientNetImageProcessor.__call__`] for details. + return_loss (`bool`, *optional*): + Whether or not to return the contrastive loss. + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + + +@dataclass +class AlignVisionModelOutput(ModelOutput): + """ + Base class for vision model's outputs that also contains image embeddings of the pooling of the last hidden states. + + Args: + image_embeds (`torch.FloatTensor` of shape `(batch_size, output_dim)` *optional* returned when model is initialized with `with_projection=True`): + The image embeddings obtained by applying the projection layer to the pooler_output. + last_hidden_state (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`): + Sequence of hidden-states at the output of the last layer of the model. + hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings, if the model has an embedding layer, + + one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`. + + Hidden-states of the model at the output of each layer plus the optional initial embedding outputs. + """ + + image_embeds: Optional[torch.FloatTensor] = None + last_hidden_state: torch.FloatTensor = None + hidden_states: Optional[Tuple[torch.FloatTensor]] = None + + +@dataclass +class AlignTextModelOutput(ModelOutput): + """ + Base class for text model's outputs that also contains a pooling of the last hidden states. + + Args: + text_embeds (`torch.FloatTensor` of shape `(batch_size, output_dim)` *optional* returned when model is initialized with `with_projection=True`): + The text embeddings obtained by applying the projection layer to the pooler_output. + last_hidden_state (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`): + Sequence of hidden-states at the output of the last layer of the model. + hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings, if the model has an embedding layer, + + one for the output of each layer) of shape `(batch_size, sequence_length, hidden_size)`. + + Hidden-states of the model at the output of each layer plus the optional initial embedding outputs. + attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, + sequence_length)`. + + Attentions weights after the attention softmax, used to compute the weighted average in the self-attention + heads. + """ + + text_embeds: Optional[torch.FloatTensor] = None + last_hidden_state: torch.FloatTensor = None + hidden_states: Optional[Tuple[torch.FloatTensor]] = None + attentions: Optional[Tuple[torch.FloatTensor]] = None + + +@dataclass +class AlignOutput(ModelOutput): + """ + Args: + loss (`torch.FloatTensor` of shape `(1,)`, *optional*, returned when `return_loss` is `True`): + Contrastive loss for image-text similarity. + logits_per_image:(`torch.FloatTensor` of shape `(image_batch_size, text_batch_size)`): + The scaled dot product scores between `image_embeds` and `text_embeds`. This represents the image-text + similarity scores. + logits_per_text:(`torch.FloatTensor` of shape `(text_batch_size, image_batch_size)`): + The scaled dot product scores between `text_embeds` and `image_embeds`. This represents the text-image + similarity scores. + text_embeds(`torch.FloatTensor` of shape `(batch_size, output_dim`): + The text embeddings obtained by applying the projection layer to the pooled output of [`AlignTextModel`]. + image_embeds(`torch.FloatTensor` of shape `(batch_size, output_dim`): + The output of [`AlignVisionModel`]. + text_model_output(`BaseModelOutputWithPoolingAndCrossAttentions`): + The output of the [`AlignTextModel`]. + vision_model_output(`BaseModelOutputWithPoolingAndNoAttention`): + The output of the [`AlignVisionModel`]. + """ + + loss: Optional[torch.FloatTensor] = None + logits_per_image: torch.FloatTensor = None + logits_per_text: torch.FloatTensor = None + text_embeds: torch.FloatTensor = None + image_embeds: torch.FloatTensor = None + text_model_output: BaseModelOutputWithPoolingAndCrossAttentions = None + vision_model_output: BaseModelOutputWithPoolingAndNoAttention = None + + def to_tuple(self) -> Tuple[Any]: + return tuple( + self[k] if k not in ["text_model_output", "vision_model_output"] else getattr(self, k).to_tuple() + for k in self.keys() + ) + + +# contrastive loss function, adapted from +# https://sachinruk.github.io/blog/pytorch/pytorch%20lightning/loss%20function/gpu/2021/03/07/CLIP.html +def contrastive_loss(logits: torch.Tensor) -> torch.Tensor: + return nn.functional.cross_entropy(logits, torch.arange(len(logits), device=logits.device), label_smoothing=0.1) + + +def align_loss(similarity: torch.Tensor) -> torch.Tensor: + caption_loss = contrastive_loss(similarity) + image_loss = contrastive_loss(similarity.t()) + return (caption_loss + image_loss) / 2.0 + + +# Copied from transformers.models.efficientnet.modeling_efficientnet.round_filters with EfficientNet->AlignVision +def round_filters(config: AlignVisionConfig, num_channels: int): + r""" + Round number of filters based on depth multiplier. + """ + divisor = config.depth_divisor + num_channels *= config.width_coefficient + new_dim = max(divisor, int(num_channels + divisor / 2) // divisor * divisor) + + # Make sure that round down does not go down by more than 10%. + if new_dim < 0.9 * num_channels: + new_dim += divisor + + return int(new_dim) + + +# Copied from transformers.models.efficientnet.modeling_efficientnet.correct_pad +def correct_pad(kernel_size: Union[int, Tuple], adjust: bool = True): + r""" + Utility function to get the tuple padding value for the depthwise convolution. + + Args: + kernel_size (`int` or `tuple`): + Kernel size of the convolution layers. + adjust (`bool`, *optional*, defaults to `True`): + Adjusts padding value to apply to right and bottom sides of the input. + """ + if isinstance(kernel_size, int): + kernel_size = (kernel_size, kernel_size) + + correct = (kernel_size[0] // 2, kernel_size[1] // 2) + if adjust: + return (correct[1] - 1, correct[1], correct[0] - 1, correct[0]) + else: + return (correct[1], correct[1], correct[0], correct[0]) + + +# Copied from transformers.models.efficientnet.modeling_efficientnet.EfficientNetEmbeddings with EfficientNet->AlignVision +class AlignVisionEmbeddings(nn.Module): + r""" + A module that corresponds to the stem module of the original work. + """ + + def __init__(self, config: AlignVisionConfig): + super().__init__() + + self.out_dim = round_filters(config, 32) + self.padding = nn.ZeroPad2d(padding=(0, 1, 0, 1)) + self.convolution = nn.Conv2d( + config.num_channels, self.out_dim, kernel_size=3, stride=2, padding="valid", bias=False + ) + self.batchnorm = nn.BatchNorm2d(self.out_dim, eps=config.batch_norm_eps, momentum=config.batch_norm_momentum) + self.activation = ACT2FN[config.hidden_act] + + def forward(self, pixel_values: torch.Tensor) -> torch.Tensor: + features = self.padding(pixel_values) + features = self.convolution(features) + features = self.batchnorm(features) + features = self.activation(features) + + return features + + +# Copied from transformers.models.efficientnet.modeling_efficientnet.EfficientNetDepthwiseConv2d with EfficientNet->AlignVision +class AlignVisionDepthwiseConv2d(nn.Conv2d): + def __init__( + self, + in_channels, + depth_multiplier=1, + kernel_size=3, + stride=1, + padding=0, + dilation=1, + bias=True, + padding_mode="zeros", + ): + out_channels = in_channels * depth_multiplier + super().__init__( + in_channels=in_channels, + out_channels=out_channels, + kernel_size=kernel_size, + stride=stride, + padding=padding, + dilation=dilation, + groups=in_channels, + bias=bias, + padding_mode=padding_mode, + ) + + +# Copied from transformers.models.efficientnet.modeling_efficientnet.EfficientNetExpansionLayer with EfficientNet->AlignVision +class AlignVisionExpansionLayer(nn.Module): + r""" + This corresponds to the expansion phase of each block in the original implementation. + """ + + def __init__(self, config: AlignVisionConfig, in_dim: int, out_dim: int, stride: int): + super().__init__() + self.expand_conv = nn.Conv2d( + in_channels=in_dim, + out_channels=out_dim, + kernel_size=1, + padding="same", + bias=False, + ) + self.expand_bn = nn.BatchNorm2d(num_features=out_dim, eps=config.batch_norm_eps) + self.expand_act = ACT2FN[config.hidden_act] + + def forward(self, hidden_states: torch.FloatTensor) -> torch.Tensor: + # Expand phase + hidden_states = self.expand_conv(hidden_states) + hidden_states = self.expand_bn(hidden_states) + hidden_states = self.expand_act(hidden_states) + + return hidden_states + + +# Copied from transformers.models.efficientnet.modeling_efficientnet.EfficientNetDepthwiseLayer with EfficientNet->AlignVision +class AlignVisionDepthwiseLayer(nn.Module): + r""" + This corresponds to the depthwise convolution phase of each block in the original implementation. + """ + + def __init__( + self, + config: AlignVisionConfig, + in_dim: int, + stride: int, + kernel_size: int, + adjust_padding: bool, + ): + super().__init__() + self.stride = stride + conv_pad = "valid" if self.stride == 2 else "same" + padding = correct_pad(kernel_size, adjust=adjust_padding) + + self.depthwise_conv_pad = nn.ZeroPad2d(padding=padding) + self.depthwise_conv = AlignVisionDepthwiseConv2d( + in_dim, kernel_size=kernel_size, stride=stride, padding=conv_pad, bias=False + ) + self.depthwise_norm = nn.BatchNorm2d( + num_features=in_dim, eps=config.batch_norm_eps, momentum=config.batch_norm_momentum + ) + self.depthwise_act = ACT2FN[config.hidden_act] + + def forward(self, hidden_states: torch.FloatTensor) -> torch.Tensor: + # Depthwise convolution + if self.stride == 2: + hidden_states = self.depthwise_conv_pad(hidden_states) + + hidden_states = self.depthwise_conv(hidden_states) + hidden_states = self.depthwise_norm(hidden_states) + hidden_states = self.depthwise_act(hidden_states) + + return hidden_states + + +# Copied from transformers.models.efficientnet.modeling_efficientnet.EfficientNetSqueezeExciteLayer with EfficientNet->AlignVision +class AlignVisionSqueezeExciteLayer(nn.Module): + r""" + This corresponds to the Squeeze and Excitement phase of each block in the original implementation. + """ + + def __init__(self, config: AlignVisionConfig, in_dim: int, expand_dim: int, expand: bool = False): + super().__init__() + self.dim = expand_dim if expand else in_dim + self.dim_se = max(1, int(in_dim * config.squeeze_expansion_ratio)) + + self.squeeze = nn.AdaptiveAvgPool2d(output_size=1) + self.reduce = nn.Conv2d( + in_channels=self.dim, + out_channels=self.dim_se, + kernel_size=1, + padding="same", + ) + self.expand = nn.Conv2d( + in_channels=self.dim_se, + out_channels=self.dim, + kernel_size=1, + padding="same", + ) + self.act_reduce = ACT2FN[config.hidden_act] + self.act_expand = nn.Sigmoid() + + def forward(self, hidden_states: torch.FloatTensor) -> torch.Tensor: + inputs = hidden_states + hidden_states = self.squeeze(hidden_states) + hidden_states = self.reduce(hidden_states) + hidden_states = self.act_reduce(hidden_states) + + hidden_states = self.expand(hidden_states) + hidden_states = self.act_expand(hidden_states) + hidden_states = torch.mul(inputs, hidden_states) + + return hidden_states + + +class AlignVisionFinalBlockLayer(nn.Module): + r""" + This corresponds to the final phase of each block in the original implementation. + """ + + def __init__( + self, config: AlignVisionConfig, in_dim: int, out_dim: int, stride: int, drop_rate: float, id_skip: bool + ): + super().__init__() + self.apply_dropout = stride == 1 and not id_skip + self.project_conv = nn.Conv2d( + in_channels=in_dim, + out_channels=out_dim, + kernel_size=1, + padding="same", + bias=False, + ) + self.project_bn = nn.BatchNorm2d( + num_features=out_dim, eps=config.batch_norm_eps, momentum=config.batch_norm_momentum + ) + self.dropout = nn.Dropout(p=drop_rate) + + def forward(self, embeddings: torch.FloatTensor, hidden_states: torch.FloatTensor) -> torch.Tensor: + hidden_states = self.project_conv(hidden_states) + hidden_states = self.project_bn(hidden_states) + + if self.apply_dropout: + hidden_states = self.dropout(hidden_states) + hidden_states = hidden_states + embeddings + + return hidden_states + + +class AlignVisionBlock(nn.Module): + r""" + This corresponds to the block module of original the EfficientNet vision encoder implementation. + + Args: + config ([`AlignVisionConfig`]): + Model configuration class. + in_dim (`int`): + Number of input channels. + out_dim (`int`): + Number of output channels. + stride (`int`): + Stride size to be used in convolution layers. + expand_ratio (`int`): + Expand ratio to set the output dimensions for the expansion and squeeze-excite layers. + kernel_size (`int`): + Kernel size for the depthwise convolution layer. + drop_rate (`float`): + Dropout rate to be used in the final phase of each block. + id_skip (`bool`): + Whether to apply dropout and sum the final hidden states with the input embeddings during the final phase + of each block. Set to `True` for the first block of each stage. + adjust_padding (`bool`): + Whether to apply padding to only right and bottom side of the input kernel before the depthwise convolution + operation, set to `True` for inputs with odd input sizes. + """ + + def __init__( + self, + config: AlignVisionConfig, + in_dim: int, + out_dim: int, + stride: int, + expand_ratio: int, + kernel_size: int, + drop_rate: float, + id_skip: bool, + adjust_padding: bool, + ): + super().__init__() + self.expand_ratio = expand_ratio + self.expand = True if self.expand_ratio != 1 else False + expand_in_dim = in_dim * expand_ratio + + if self.expand: + self.expansion = AlignVisionExpansionLayer( + config=config, in_dim=in_dim, out_dim=expand_in_dim, stride=stride + ) + + self.depthwise_conv = AlignVisionDepthwiseLayer( + config=config, + in_dim=expand_in_dim if self.expand else in_dim, + stride=stride, + kernel_size=kernel_size, + adjust_padding=adjust_padding, + ) + self.squeeze_excite = AlignVisionSqueezeExciteLayer( + config=config, in_dim=in_dim, expand_dim=expand_in_dim, expand=self.expand + ) + self.projection = AlignVisionFinalBlockLayer( + config=config, + in_dim=expand_in_dim if self.expand else in_dim, + out_dim=out_dim, + stride=stride, + drop_rate=drop_rate, + id_skip=id_skip, + ) + + def forward(self, hidden_states: torch.FloatTensor) -> torch.Tensor: + embeddings = hidden_states + # Expansion and depthwise convolution phase + if self.expand_ratio != 1: + hidden_states = self.expansion(hidden_states) + hidden_states = self.depthwise_conv(hidden_states) + + # Squeeze and excite phase + hidden_states = self.squeeze_excite(hidden_states) + hidden_states = self.projection(embeddings, hidden_states) + return hidden_states + + +class AlignVisionEncoder(nn.Module): + r""" + Forward propogates the embeddings through each vision encoder (EfficientNet) block. + + Args: + config ([`AlignVisionConfig`]): + Model configuration class. + """ + + def __init__(self, config: AlignVisionConfig): + super().__init__() + self.depth_coefficient = config.depth_coefficient + + def round_repeats(repeats): + # Round number of block repeats based on depth multiplier. + return int(math.ceil(self.depth_coefficient * repeats)) + + num_base_blocks = len(config.in_channels) + num_blocks = sum(round_repeats(n) for n in config.num_block_repeats) + + curr_block_num = 0 + blocks = [] + for i in range(num_base_blocks): + in_dim = round_filters(config, config.in_channels[i]) + out_dim = round_filters(config, config.out_channels[i]) + stride = config.strides[i] + kernel_size = config.kernel_sizes[i] + expand_ratio = config.expand_ratios[i] + + for j in range(round_repeats(config.num_block_repeats[i])): + id_skip = True if j == 0 else False + stride = 1 if j > 0 else stride + in_dim = out_dim if j > 0 else in_dim + adjust_padding = False if curr_block_num in config.depthwise_padding else True + drop_rate = config.drop_connect_rate * curr_block_num / num_blocks + + block = AlignVisionBlock( + config=config, + in_dim=in_dim, + out_dim=out_dim, + stride=stride, + kernel_size=kernel_size, + expand_ratio=expand_ratio, + drop_rate=drop_rate, + id_skip=id_skip, + adjust_padding=adjust_padding, + ) + blocks.append(block) + curr_block_num += 1 + + self.blocks = nn.ModuleList(blocks) + + def forward( + self, + hidden_states: torch.FloatTensor, + output_hidden_states: Optional[bool] = False, + return_dict: Optional[bool] = True, + ) -> BaseModelOutputWithPoolingAndNoAttention: + all_hidden_states = (hidden_states,) if output_hidden_states else None + + for block in self.blocks: + hidden_states = block(hidden_states) + if output_hidden_states: + all_hidden_states += (hidden_states,) + + if not return_dict: + return tuple(v for v in [hidden_states, all_hidden_states] if v is not None) + + return BaseModelOutputWithNoAttention( + last_hidden_state=hidden_states, + hidden_states=all_hidden_states, + ) + + +# Copied from transformers.models.bert.modeling_bert.BertEmbeddings with Bert->AlignText +class AlignTextEmbeddings(nn.Module): + """Construct the embeddings from word, position and token_type embeddings.""" + + def __init__(self, config): + super().__init__() + self.word_embeddings = nn.Embedding(config.vocab_size, config.hidden_size, padding_idx=config.pad_token_id) + self.position_embeddings = nn.Embedding(config.max_position_embeddings, config.hidden_size) + self.token_type_embeddings = nn.Embedding(config.type_vocab_size, config.hidden_size) + + # self.LayerNorm is not snake-cased to stick with TensorFlow model variable name and be able to load + # any TensorFlow checkpoint file + self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + # position_ids (1, len position emb) is contiguous in memory and exported when serialized + self.position_embedding_type = getattr(config, "position_embedding_type", "absolute") + self.register_buffer( + "position_ids", torch.arange(config.max_position_embeddings).expand((1, -1)), persistent=False + ) + self.register_buffer( + "token_type_ids", torch.zeros(self.position_ids.size(), dtype=torch.long), persistent=False + ) + + def forward( + self, + input_ids: Optional[torch.LongTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.LongTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + past_key_values_length: int = 0, + ) -> torch.Tensor: + if input_ids is not None: + input_shape = input_ids.size() + else: + input_shape = inputs_embeds.size()[:-1] + + seq_length = input_shape[1] + + if position_ids is None: + position_ids = self.position_ids[:, past_key_values_length : seq_length + past_key_values_length] + + # Setting the token_type_ids to the registered buffer in constructor where it is all zeros, which usually occurs + # when its auto-generated, registered buffer helps users when tracing the model without passing token_type_ids, solves + # issue #5664 + if token_type_ids is None: + if hasattr(self, "token_type_ids"): + buffered_token_type_ids = self.token_type_ids[:, :seq_length] + buffered_token_type_ids_expanded = buffered_token_type_ids.expand(input_shape[0], seq_length) + token_type_ids = buffered_token_type_ids_expanded + else: + token_type_ids = torch.zeros(input_shape, dtype=torch.long, device=self.position_ids.device) + + if inputs_embeds is None: + inputs_embeds = self.word_embeddings(input_ids) + token_type_embeddings = self.token_type_embeddings(token_type_ids) + + embeddings = inputs_embeds + token_type_embeddings + if self.position_embedding_type == "absolute": + position_embeddings = self.position_embeddings(position_ids) + embeddings += position_embeddings + embeddings = self.LayerNorm(embeddings) + embeddings = self.dropout(embeddings) + return embeddings + + +# Copied from transformers.models.bert.modeling_bert.BertSelfAttention with Bert->AlignText +class AlignTextSelfAttention(nn.Module): + def __init__(self, config, position_embedding_type=None): + super().__init__() + if config.hidden_size % config.num_attention_heads != 0 and not hasattr(config, "embedding_size"): + raise ValueError( + f"The hidden size ({config.hidden_size}) is not a multiple of the number of attention " + f"heads ({config.num_attention_heads})" + ) + + self.num_attention_heads = config.num_attention_heads + self.attention_head_size = int(config.hidden_size / config.num_attention_heads) + self.all_head_size = self.num_attention_heads * self.attention_head_size + + self.query = nn.Linear(config.hidden_size, self.all_head_size) + self.key = nn.Linear(config.hidden_size, self.all_head_size) + self.value = nn.Linear(config.hidden_size, self.all_head_size) + + self.dropout = nn.Dropout(config.attention_probs_dropout_prob) + self.position_embedding_type = position_embedding_type or getattr( + config, "position_embedding_type", "absolute" + ) + if self.position_embedding_type == "relative_key" or self.position_embedding_type == "relative_key_query": + self.max_position_embeddings = config.max_position_embeddings + self.distance_embedding = nn.Embedding(2 * config.max_position_embeddings - 1, self.attention_head_size) + + self.is_decoder = config.is_decoder + + def transpose_for_scores(self, x: torch.Tensor) -> torch.Tensor: + new_x_shape = x.size()[:-1] + (self.num_attention_heads, self.attention_head_size) + x = x.view(new_x_shape) + return x.permute(0, 2, 1, 3) + + def forward( + self, + hidden_states: torch.Tensor, + attention_mask: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_value: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + output_attentions: Optional[bool] = False, + ) -> Tuple[torch.Tensor]: + mixed_query_layer = self.query(hidden_states) + + # If this is instantiated as a cross-attention module, the keys + # and values come from an encoder; the attention mask needs to be + # such that the encoder's padding tokens are not attended to. + is_cross_attention = encoder_hidden_states is not None + + if is_cross_attention and past_key_value is not None: + # reuse k,v, cross_attentions + key_layer = past_key_value[0] + value_layer = past_key_value[1] + attention_mask = encoder_attention_mask + elif is_cross_attention: + key_layer = self.transpose_for_scores(self.key(encoder_hidden_states)) + value_layer = self.transpose_for_scores(self.value(encoder_hidden_states)) + attention_mask = encoder_attention_mask + elif past_key_value is not None: + key_layer = self.transpose_for_scores(self.key(hidden_states)) + value_layer = self.transpose_for_scores(self.value(hidden_states)) + key_layer = torch.cat([past_key_value[0], key_layer], dim=2) + value_layer = torch.cat([past_key_value[1], value_layer], dim=2) + else: + key_layer = self.transpose_for_scores(self.key(hidden_states)) + value_layer = self.transpose_for_scores(self.value(hidden_states)) + + query_layer = self.transpose_for_scores(mixed_query_layer) + + use_cache = past_key_value is not None + if self.is_decoder: + # if cross_attention save Tuple(torch.Tensor, torch.Tensor) of all cross attention key/value_states. + # Further calls to cross_attention layer can then reuse all cross-attention + # key/value_states (first "if" case) + # if uni-directional self-attention (decoder) save Tuple(torch.Tensor, torch.Tensor) of + # all previous decoder key/value_states. Further calls to uni-directional self-attention + # can concat previous decoder key/value_states to current projected key/value_states (third "elif" case) + # if encoder bi-directional self-attention `past_key_value` is always `None` + past_key_value = (key_layer, value_layer) + + # Take the dot product between "query" and "key" to get the raw attention scores. + attention_scores = torch.matmul(query_layer, key_layer.transpose(-1, -2)) + + if self.position_embedding_type == "relative_key" or self.position_embedding_type == "relative_key_query": + query_length, key_length = query_layer.shape[2], key_layer.shape[2] + if use_cache: + position_ids_l = torch.tensor(key_length - 1, dtype=torch.long, device=hidden_states.device).view( + -1, 1 + ) + else: + position_ids_l = torch.arange(query_length, dtype=torch.long, device=hidden_states.device).view(-1, 1) + position_ids_r = torch.arange(key_length, dtype=torch.long, device=hidden_states.device).view(1, -1) + distance = position_ids_l - position_ids_r + + positional_embedding = self.distance_embedding(distance + self.max_position_embeddings - 1) + positional_embedding = positional_embedding.to(dtype=query_layer.dtype) # fp16 compatibility + + if self.position_embedding_type == "relative_key": + relative_position_scores = torch.einsum("bhld,lrd->bhlr", query_layer, positional_embedding) + attention_scores = attention_scores + relative_position_scores + elif self.position_embedding_type == "relative_key_query": + relative_position_scores_query = torch.einsum("bhld,lrd->bhlr", query_layer, positional_embedding) + relative_position_scores_key = torch.einsum("bhrd,lrd->bhlr", key_layer, positional_embedding) + attention_scores = attention_scores + relative_position_scores_query + relative_position_scores_key + + attention_scores = attention_scores / math.sqrt(self.attention_head_size) + if attention_mask is not None: + # Apply the attention mask is (precomputed for all layers in AlignTextModel forward() function) + attention_scores = attention_scores + attention_mask + + # Normalize the attention scores to probabilities. + attention_probs = nn.functional.softmax(attention_scores, dim=-1) + + # This is actually dropping out entire tokens to attend to, which might + # seem a bit unusual, but is taken from the original Transformer paper. + attention_probs = self.dropout(attention_probs) + + # Mask heads if we want to + if head_mask is not None: + attention_probs = attention_probs * head_mask + + context_layer = torch.matmul(attention_probs, value_layer) + + context_layer = context_layer.permute(0, 2, 1, 3).contiguous() + new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) + context_layer = context_layer.view(new_context_layer_shape) + + outputs = (context_layer, attention_probs) if output_attentions else (context_layer,) + + if self.is_decoder: + outputs = outputs + (past_key_value,) + return outputs + + +# Copied from transformers.models.bert.modeling_bert.BertSelfOutput with Bert->AlignText +class AlignTextSelfOutput(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.hidden_size) + self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.dropout(hidden_states) + hidden_states = self.LayerNorm(hidden_states + input_tensor) + return hidden_states + + +# Copied from transformers.models.bert.modeling_bert.BertAttention with Bert->AlignText +class AlignTextAttention(nn.Module): + def __init__(self, config, position_embedding_type=None): + super().__init__() + self.self = AlignTextSelfAttention(config, position_embedding_type=position_embedding_type) + self.output = AlignTextSelfOutput(config) + self.pruned_heads = set() + + def prune_heads(self, heads): + if len(heads) == 0: + return + heads, index = find_pruneable_heads_and_indices( + heads, self.self.num_attention_heads, self.self.attention_head_size, self.pruned_heads + ) + + # Prune linear layers + self.self.query = prune_linear_layer(self.self.query, index) + self.self.key = prune_linear_layer(self.self.key, index) + self.self.value = prune_linear_layer(self.self.value, index) + self.output.dense = prune_linear_layer(self.output.dense, index, dim=1) + + # Update hyper params and store pruned heads + self.self.num_attention_heads = self.self.num_attention_heads - len(heads) + self.self.all_head_size = self.self.attention_head_size * self.self.num_attention_heads + self.pruned_heads = self.pruned_heads.union(heads) + + def forward( + self, + hidden_states: torch.Tensor, + attention_mask: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_value: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + output_attentions: Optional[bool] = False, + ) -> Tuple[torch.Tensor]: + self_outputs = self.self( + hidden_states, + attention_mask, + head_mask, + encoder_hidden_states, + encoder_attention_mask, + past_key_value, + output_attentions, + ) + attention_output = self.output(self_outputs[0], hidden_states) + outputs = (attention_output,) + self_outputs[1:] # add attentions if we output them + return outputs + + +# Copied from transformers.models.bert.modeling_bert.BertIntermediate with Bert->AlignText +class AlignTextIntermediate(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.intermediate_size) + if isinstance(config.hidden_act, str): + self.intermediate_act_fn = ACT2FN[config.hidden_act] + else: + self.intermediate_act_fn = config.hidden_act + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.intermediate_act_fn(hidden_states) + return hidden_states + + +# Copied from transformers.models.bert.modeling_bert.BertOutput with Bert->AlignText +class AlignTextOutput(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.intermediate_size, config.hidden_size) + self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.dropout(hidden_states) + hidden_states = self.LayerNorm(hidden_states + input_tensor) + return hidden_states + + +# Copied from transformers.models.bert.modeling_bert.BertLayer with Bert->AlignText +class AlignTextLayer(nn.Module): + def __init__(self, config): + super().__init__() + self.chunk_size_feed_forward = config.chunk_size_feed_forward + self.seq_len_dim = 1 + self.attention = AlignTextAttention(config) + self.is_decoder = config.is_decoder + self.add_cross_attention = config.add_cross_attention + if self.add_cross_attention: + if not self.is_decoder: + raise ValueError(f"{self} should be used as a decoder model if cross attention is added") + self.crossattention = AlignTextAttention(config, position_embedding_type="absolute") + self.intermediate = AlignTextIntermediate(config) + self.output = AlignTextOutput(config) + + def forward( + self, + hidden_states: torch.Tensor, + attention_mask: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_value: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + output_attentions: Optional[bool] = False, + ) -> Tuple[torch.Tensor]: + # decoder uni-directional self-attention cached key/values tuple is at positions 1,2 + self_attn_past_key_value = past_key_value[:2] if past_key_value is not None else None + self_attention_outputs = self.attention( + hidden_states, + attention_mask, + head_mask, + output_attentions=output_attentions, + past_key_value=self_attn_past_key_value, + ) + attention_output = self_attention_outputs[0] + + # if decoder, the last output is tuple of self-attn cache + if self.is_decoder: + outputs = self_attention_outputs[1:-1] + present_key_value = self_attention_outputs[-1] + else: + outputs = self_attention_outputs[1:] # add self attentions if we output attention weights + + cross_attn_present_key_value = None + if self.is_decoder and encoder_hidden_states is not None: + if not hasattr(self, "crossattention"): + raise ValueError( + f"If `encoder_hidden_states` are passed, {self} has to be instantiated with cross-attention layers" + " by setting `config.add_cross_attention=True`" + ) + + # cross_attn cached key/values tuple is at positions 3,4 of past_key_value tuple + cross_attn_past_key_value = past_key_value[-2:] if past_key_value is not None else None + cross_attention_outputs = self.crossattention( + attention_output, + attention_mask, + head_mask, + encoder_hidden_states, + encoder_attention_mask, + cross_attn_past_key_value, + output_attentions, + ) + attention_output = cross_attention_outputs[0] + outputs = outputs + cross_attention_outputs[1:-1] # add cross attentions if we output attention weights + + # add cross-attn cache to positions 3,4 of present_key_value tuple + cross_attn_present_key_value = cross_attention_outputs[-1] + present_key_value = present_key_value + cross_attn_present_key_value + + layer_output = apply_chunking_to_forward( + self.feed_forward_chunk, self.chunk_size_feed_forward, self.seq_len_dim, attention_output + ) + outputs = (layer_output,) + outputs + + # if decoder, return the attn key/values as the last output + if self.is_decoder: + outputs = outputs + (present_key_value,) + + return outputs + + def feed_forward_chunk(self, attention_output): + intermediate_output = self.intermediate(attention_output) + layer_output = self.output(intermediate_output, attention_output) + return layer_output + + +# Copied from transformers.models.bert.modeling_bert.BertEncoder with Bert->AlignText +class AlignTextEncoder(nn.Module): + def __init__(self, config): + super().__init__() + self.config = config + self.layer = nn.ModuleList([AlignTextLayer(config) for _ in range(config.num_hidden_layers)]) + self.gradient_checkpointing = False + + def forward( + self, + hidden_states: torch.Tensor, + attention_mask: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = False, + output_hidden_states: Optional[bool] = False, + return_dict: Optional[bool] = True, + ) -> Union[Tuple[torch.Tensor], BaseModelOutputWithPastAndCrossAttentions]: + all_hidden_states = () if output_hidden_states else None + all_self_attentions = () if output_attentions else None + all_cross_attentions = () if output_attentions and self.config.add_cross_attention else None + + if self.gradient_checkpointing and self.training: + if use_cache: + logger.warning_once( + "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..." + ) + use_cache = False + + next_decoder_cache = () if use_cache else None + for i, layer_module in enumerate(self.layer): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + layer_head_mask = head_mask[i] if head_mask is not None else None + past_key_value = past_key_values[i] if past_key_values is not None else None + + if self.gradient_checkpointing and self.training: + layer_outputs = self._gradient_checkpointing_func( + layer_module.__call__, + hidden_states, + attention_mask, + layer_head_mask, + encoder_hidden_states, + encoder_attention_mask, + past_key_value, + output_attentions, + ) + else: + layer_outputs = layer_module( + hidden_states, + attention_mask, + layer_head_mask, + encoder_hidden_states, + encoder_attention_mask, + past_key_value, + output_attentions, + ) + + hidden_states = layer_outputs[0] + if use_cache: + next_decoder_cache += (layer_outputs[-1],) + if output_attentions: + all_self_attentions = all_self_attentions + (layer_outputs[1],) + if self.config.add_cross_attention: + all_cross_attentions = all_cross_attentions + (layer_outputs[2],) + + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if not return_dict: + return tuple( + v + for v in [ + hidden_states, + next_decoder_cache, + all_hidden_states, + all_self_attentions, + all_cross_attentions, + ] + if v is not None + ) + return BaseModelOutputWithPastAndCrossAttentions( + last_hidden_state=hidden_states, + past_key_values=next_decoder_cache, + hidden_states=all_hidden_states, + attentions=all_self_attentions, + cross_attentions=all_cross_attentions, + ) + + +# Copied from transformers.models.bert.modeling_bert.BertPooler with Bert -> AlignText +class AlignTextPooler(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.hidden_size) + self.activation = nn.Tanh() + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + # We "pool" the model by simply taking the hidden state corresponding + # to the first token. + first_token_tensor = hidden_states[:, 0] + pooled_output = self.dense(first_token_tensor) + pooled_output = self.activation(pooled_output) + return pooled_output + + +class AlignPreTrainedModel(PreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = AlignConfig + base_model_prefix = "align" + supports_gradient_checkpointing = True + + def _init_weights(self, module): + """Initialize the weights""" + if isinstance(module, (nn.Linear, nn.Conv2d)): + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.bias is not None: + module.bias.data.zero_() + elif isinstance(module, AlignModel): + nn.init.xavier_uniform_(module.text_projection.weight) + module.text_projection.bias.data.zero_() + module.text_projection._is_hf_initialized = True + elif isinstance(module, nn.Embedding): + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.padding_idx is not None: + module.weight.data[module.padding_idx].zero_() + if isinstance(module, nn.LayerNorm): + module.bias.data.zero_() + module.weight.data.fill_(1.0) + + +@add_start_docstrings( + """The text model from ALIGN without any head or projection on top.""", + ALIGN_START_DOCSTRING, +) +class AlignTextModel(AlignPreTrainedModel): + config_class = AlignTextConfig + + def __init__(self, config: AlignTextConfig, add_pooling_layer: bool = True): + super().__init__(config) + self.config = config + + self.embeddings = AlignTextEmbeddings(config) + self.encoder = AlignTextEncoder(config) + + self.pooler = AlignTextPooler(config) if add_pooling_layer else None + + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self): + return self.embeddings.word_embeddings + + def set_input_embeddings(self, value): + self.embeddings.word_embeddings = value + + @add_start_docstrings_to_model_forward(ALIGN_TEXT_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=BaseModelOutputWithPoolingAndCrossAttentions, config_class=AlignTextConfig) + def forward( + self, + input_ids: Optional[torch.Tensor] = None, + attention_mask: Optional[torch.Tensor] = None, + token_type_ids: Optional[torch.Tensor] = None, + position_ids: Optional[torch.Tensor] = None, + head_mask: Optional[torch.Tensor] = None, + inputs_embeds: Optional[torch.Tensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, BaseModelOutputWithPoolingAndCrossAttentions]: + r""" + Returns: + + Examples: + + ```python + >>> from transformers import AutoTokenizer, AlignTextModel + + >>> model = AlignTextModel.from_pretrained("kakaobrain/align-base") + >>> tokenizer = AutoTokenizer.from_pretrained("kakaobrain/align-base") + + >>> inputs = tokenizer(["a photo of a cat", "a photo of a dog"], padding=True, return_tensors="pt") + + >>> outputs = model(**inputs) + >>> last_hidden_state = outputs.last_hidden_state + >>> pooled_output = outputs.pooler_output # pooled (EOS token) states + ```""" + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if input_ids is not None and inputs_embeds is not None: + raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time") + elif input_ids is not None: + self.warn_if_padding_and_no_attention_mask(input_ids, attention_mask) + input_shape = input_ids.size() + elif inputs_embeds is not None: + input_shape = inputs_embeds.size()[:-1] + else: + raise ValueError("You have to specify either input_ids or inputs_embeds") + + batch_size, seq_length = input_shape + device = input_ids.device if input_ids is not None else inputs_embeds.device + + if attention_mask is None: + attention_mask = torch.ones(((batch_size, seq_length)), device=device) + + if token_type_ids is None: + if hasattr(self.embeddings, "token_type_ids"): + buffered_token_type_ids = self.embeddings.token_type_ids[:, :seq_length] + buffered_token_type_ids_expanded = buffered_token_type_ids.expand(batch_size, seq_length) + token_type_ids = buffered_token_type_ids_expanded + else: + token_type_ids = torch.zeros(input_shape, dtype=torch.long, device=device) + + # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] + # ourselves in which case we just need to make it broadcastable to all heads. + extended_attention_mask: torch.Tensor = self.get_extended_attention_mask(attention_mask, input_shape) + + # Prepare head mask if needed + # 1.0 in head_mask indicate we keep the head + # attention_probs has shape bsz x n_heads x N x N + # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] + # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] + head_mask = self.get_head_mask(head_mask, self.config.num_hidden_layers) + + embedding_output = self.embeddings( + input_ids=input_ids, + position_ids=position_ids, + token_type_ids=token_type_ids, + inputs_embeds=inputs_embeds, + ) + encoder_outputs = self.encoder( + embedding_output, + attention_mask=extended_attention_mask, + head_mask=head_mask, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + sequence_output = encoder_outputs[0] + pooled_output = self.pooler(sequence_output) if self.pooler is not None else None + + if not return_dict: + return (sequence_output, pooled_output) + encoder_outputs[1:] + + return BaseModelOutputWithPoolingAndCrossAttentions( + last_hidden_state=sequence_output, + pooler_output=pooled_output, + hidden_states=encoder_outputs.hidden_states, + attentions=encoder_outputs.attentions, + cross_attentions=encoder_outputs.cross_attentions, + ) + + +@add_start_docstrings( + """The vision model from ALIGN without any head or projection on top.""", + ALIGN_START_DOCSTRING, +) +class AlignVisionModel(AlignPreTrainedModel): + config_class = AlignVisionConfig + main_input_name = "pixel_values" + supports_gradient_checkpointing = False + + def __init__(self, config: AlignVisionConfig): + super().__init__(config) + self.config = config + self.embeddings = AlignVisionEmbeddings(config) + self.encoder = AlignVisionEncoder(config) + + # Final pooling layer + if config.pooling_type == "mean": + self.pooler = nn.AvgPool2d(config.hidden_dim, ceil_mode=True) + elif config.pooling_type == "max": + self.pooler = nn.MaxPool2d(config.hidden_dim, ceil_mode=True) + else: + raise ValueError(f"config.pooling must be one of ['mean', 'max'] got {config.pooling}") + + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self) -> nn.Module: + return self.vision_model.embeddings.convolution + + @add_start_docstrings_to_model_forward(ALIGN_VISION_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=BaseModelOutputWithPoolingAndNoAttention, config_class=AlignVisionConfig) + def forward( + self, + pixel_values: Optional[torch.FloatTensor] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, BaseModelOutputWithPoolingAndNoAttention]: + r""" + Returns: + + Examples: + + ```python + >>> from PIL import Image + >>> import requests + >>> from transformers import AutoProcessor, AlignVisionModel + + >>> model = AlignVisionModel.from_pretrained("kakaobrain/align-base") + >>> processor = AutoProcessor.from_pretrained("kakaobrain/align-base") + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> inputs = processor(images=image, return_tensors="pt") + + >>> outputs = model(**inputs) + >>> last_hidden_state = outputs.last_hidden_state + >>> pooled_output = outputs.pooler_output # pooled CLS states + ```""" + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if pixel_values is None: + raise ValueError("You have to specify pixel_values") + + embedding_output = self.embeddings(pixel_values) + encoder_outputs = self.encoder( + embedding_output, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + # Apply pooling + last_hidden_state = encoder_outputs[0] + pooled_output = self.pooler(last_hidden_state) + # Reshape (batch_size, projection_dim, 1 , 1) -> (batch_size, projection_dim) + pooled_output = pooled_output.reshape(pooled_output.shape[:2]) + + if not return_dict: + return (last_hidden_state, pooled_output) + encoder_outputs[1:] + + return BaseModelOutputWithPoolingAndNoAttention( + last_hidden_state=last_hidden_state, + pooler_output=pooled_output, + hidden_states=encoder_outputs.hidden_states, + ) + + +@add_start_docstrings(ALIGN_START_DOCSTRING) +class AlignModel(AlignPreTrainedModel): + config_class = AlignConfig + + def __init__(self, config: AlignConfig): + super().__init__(config) + + if not isinstance(config.text_config, AlignTextConfig): + raise ValueError( + "config.text_config is expected to be of type AlignTextConfig but is of type" + f" {type(config.text_config)}." + ) + + if not isinstance(config.vision_config, AlignVisionConfig): + raise ValueError( + "config.vision_config is expected to be of type AlignVisionConfig but is of type" + f" {type(config.vision_config)}." + ) + + text_config = config.text_config + vision_config = config.vision_config + + self.projection_dim = config.projection_dim + self.text_embed_dim = text_config.hidden_size + + self.text_model = AlignTextModel(text_config) + self.vision_model = AlignVisionModel(vision_config) + + self.text_projection = nn.Linear(self.text_embed_dim, self.projection_dim) + self.temperature = nn.Parameter(torch.tensor(self.config.temperature_init_value)) + + # Initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(ALIGN_TEXT_INPUTS_DOCSTRING) + def get_text_features( + self, + input_ids: Optional[torch.Tensor] = None, + attention_mask: Optional[torch.Tensor] = None, + token_type_ids: Optional[torch.Tensor] = None, + position_ids: Optional[torch.Tensor] = None, + head_mask: Optional[torch.Tensor] = None, + inputs_embeds: Optional[torch.Tensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> torch.FloatTensor: + r""" + Returns: + text_features (`torch.FloatTensor` of shape `(batch_size, output_dim`): The text embeddings obtained by + applying the projection layer to the pooled output of [`AlignTextModel`]. + + Examples: + + ```python + >>> from transformers import AutoTokenizer, AlignModel + + >>> model = AlignModel.from_pretrained("kakaobrain/align-base") + >>> tokenizer = AutoTokenizer.from_pretrained("kakaobrain/align-base") + + >>> inputs = tokenizer(["a photo of a cat", "a photo of a dog"], padding=True, return_tensors="pt") + >>> text_features = model.get_text_features(**inputs) + ```""" + # Use ALIGN model's config for some fields (if specified) instead of those of vision & text components. + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + text_outputs = self.text_model( + input_ids=input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + last_hidden_state = text_outputs[0][:, 0, :] + text_features = self.text_projection(last_hidden_state) + + return text_features + + @add_start_docstrings_to_model_forward(ALIGN_VISION_INPUTS_DOCSTRING) + def get_image_features( + self, + pixel_values: Optional[torch.FloatTensor] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> torch.FloatTensor: + r""" + Returns: + image_features (`torch.FloatTensor` of shape `(batch_size, output_dim`): The image embeddings obtained by + applying the projection layer to the pooled output of [`AlignVisionModel`]. + + Examples: + + ```python + >>> from PIL import Image + >>> import requests + >>> from transformers import AutoProcessor, AlignModel + + >>> model = AlignModel.from_pretrained("kakaobrain/align-base") + >>> processor = AutoProcessor.from_pretrained("kakaobrain/align-base") + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> inputs = processor(images=image, return_tensors="pt") + + >>> image_features = model.get_image_features(**inputs) + ```""" + # Use ALIGN model's config for some fields (if specified) instead of those of vision & text components. + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + vision_outputs = self.vision_model( + pixel_values=pixel_values, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + image_features = vision_outputs[1] # pooled_output + + return image_features + + @add_start_docstrings_to_model_forward(ALIGN_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=AlignOutput, config_class=AlignConfig) + def forward( + self, + input_ids: Optional[torch.LongTensor] = None, + pixel_values: Optional[torch.FloatTensor] = None, + attention_mask: Optional[torch.Tensor] = None, + token_type_ids: Optional[torch.Tensor] = None, + position_ids: Optional[torch.Tensor] = None, + head_mask: Optional[torch.Tensor] = None, + inputs_embeds: Optional[torch.Tensor] = None, + return_loss: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, AlignOutput]: + r""" + Returns: + + Examples: + + ```python + >>> from PIL import Image + >>> import requests + >>> from transformers import AutoProcessor, AlignModel + + >>> model = AlignModel.from_pretrained("kakaobrain/align-base") + >>> processor = AutoProcessor.from_pretrained("kakaobrain/align-base") + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> inputs = processor( + ... text=["a photo of a cat", "a photo of a dog"], images=image, return_tensors="pt", padding=True + ... ) + + >>> outputs = model(**inputs) + >>> logits_per_image = outputs.logits_per_image # this is the image-text similarity score + >>> probs = logits_per_image.softmax(dim=1) # we can take the softmax to get the label probabilities + ```""" + # Use ALIGN model's config for some fields (if specified) instead of those of vision & text components. + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + vision_outputs = self.vision_model( + pixel_values=pixel_values, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + text_outputs = self.text_model( + input_ids=input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + image_embeds = vision_outputs[1] + text_embeds = text_outputs[0][:, 0, :] + text_embeds = self.text_projection(text_embeds) + + # normalized features + image_embeds = image_embeds / image_embeds.norm(p=2, dim=-1, keepdim=True) + text_embeds = text_embeds / text_embeds.norm(p=2, dim=-1, keepdim=True) + + # cosine similarity as logits + logits_per_text = torch.matmul(text_embeds, image_embeds.t()) / self.temperature + logits_per_image = logits_per_text.t() + + loss = None + if return_loss: + loss = align_loss(logits_per_text) + + if not return_dict: + output = (logits_per_image, logits_per_text, text_embeds, image_embeds, text_outputs, vision_outputs) + return ((loss,) + output) if loss is not None else output + + return AlignOutput( + loss=loss, + logits_per_image=logits_per_image, + logits_per_text=logits_per_text, + text_embeds=text_embeds, + image_embeds=image_embeds, + text_model_output=text_outputs, + vision_model_output=vision_outputs, + ) diff --git a/venv/lib/python3.10/site-packages/transformers/models/align/processing_align.py b/venv/lib/python3.10/site-packages/transformers/models/align/processing_align.py new file mode 100644 index 0000000000000000000000000000000000000000..8bcea7eb5dadf61d0263bd60aea328b7da8714c9 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/align/processing_align.py @@ -0,0 +1,121 @@ +# coding=utf-8 +# Copyright 2023 The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" +Image/Text processor class for ALIGN +""" + + +from ...processing_utils import ProcessorMixin +from ...tokenization_utils_base import BatchEncoding + + +class AlignProcessor(ProcessorMixin): + r""" + Constructs an ALIGN processor which wraps [`EfficientNetImageProcessor`] and + [`BertTokenizer`]/[`BertTokenizerFast`] into a single processor that interits both the image processor and + tokenizer functionalities. See the [`~AlignProcessor.__call__`] and [`~OwlViTProcessor.decode`] for more + information. + + Args: + image_processor ([`EfficientNetImageProcessor`]): + The image processor is a required input. + tokenizer ([`BertTokenizer`, `BertTokenizerFast`]): + The tokenizer is a required input. + """ + + attributes = ["image_processor", "tokenizer"] + image_processor_class = "EfficientNetImageProcessor" + tokenizer_class = ("BertTokenizer", "BertTokenizerFast") + + def __init__(self, image_processor, tokenizer): + super().__init__(image_processor, tokenizer) + + def __call__(self, text=None, images=None, padding="max_length", max_length=64, return_tensors=None, **kwargs): + """ + Main method to prepare text(s) and image(s) to be fed as input to the model. This method forwards the `text` + and `kwargs` arguments to BertTokenizerFast's [`~BertTokenizerFast.__call__`] if `text` is not `None` to encode + the text. To prepare the image(s), this method forwards the `images` and `kwargs` arguments to + EfficientNetImageProcessor's [`~EfficientNetImageProcessor.__call__`] if `images` is not `None`. Please refer + to the doctsring of the above two methods for more information. + + Args: + text (`str`, `List[str]`): + The sequence or batch of sequences to be encoded. Each sequence can be a string or a list of strings + (pretokenized string). If the sequences are provided as list of strings (pretokenized), you must set + `is_split_into_words=True` (to lift the ambiguity with a batch of sequences). + images (`PIL.Image.Image`, `np.ndarray`, `torch.Tensor`, `List[PIL.Image.Image]`, `List[np.ndarray]`, `List[torch.Tensor]`): + The image or batch of images to be prepared. Each image can be a PIL image, NumPy array or PyTorch + tensor. Both channels-first and channels-last formats are supported. + padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `max_length`): + Activates and controls padding for tokenization of input text. Choose between [`True` or `'longest'`, + `'max_length'`, `False` or `'do_not_pad'`] + max_length (`int`, *optional*, defaults to `max_length`): + Maximum padding value to use to pad the input text during tokenization. + + return_tensors (`str` or [`~utils.TensorType`], *optional*): + If set, will return tensors of a particular framework. Acceptable values are: + + - `'tf'`: Return TensorFlow `tf.constant` objects. + - `'pt'`: Return PyTorch `torch.Tensor` objects. + - `'np'`: Return NumPy `np.ndarray` objects. + - `'jax'`: Return JAX `jnp.ndarray` objects. + + Returns: + [`BatchEncoding`]: A [`BatchEncoding`] with the following fields: + + - **input_ids** -- List of token ids to be fed to a model. Returned when `text` is not `None`. + - **attention_mask** -- List of indices specifying which tokens should be attended to by the model (when + `return_attention_mask=True` or if *"attention_mask"* is in `self.model_input_names` and if `text` is not + `None`). + - **pixel_values** -- Pixel values to be fed to a model. Returned when `images` is not `None`. + """ + if text is None and images is None: + raise ValueError("You have to specify either text or images. Both cannot be none.") + + if text is not None: + encoding = self.tokenizer( + text, padding=padding, max_length=max_length, return_tensors=return_tensors, **kwargs + ) + + if images is not None: + image_features = self.image_processor(images, return_tensors=return_tensors, **kwargs) + + if text is not None and images is not None: + encoding["pixel_values"] = image_features.pixel_values + return encoding + elif text is not None: + return encoding + else: + return BatchEncoding(data=dict(**image_features), tensor_type=return_tensors) + + def batch_decode(self, *args, **kwargs): + """ + This method forwards all its arguments to BertTokenizerFast's [`~PreTrainedTokenizer.batch_decode`]. Please + refer to the docstring of this method for more information. + """ + return self.tokenizer.batch_decode(*args, **kwargs) + + def decode(self, *args, **kwargs): + """ + This method forwards all its arguments to BertTokenizerFast's [`~PreTrainedTokenizer.decode`]. Please refer to + the docstring of this method for more information. + """ + return self.tokenizer.decode(*args, **kwargs) + + @property + def model_input_names(self): + tokenizer_input_names = self.tokenizer.model_input_names + image_processor_input_names = self.image_processor.model_input_names + return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names)) diff --git a/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/__init__.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/__init__.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..710f4f4756ae934169f0f778c707f59d6089bd4d Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/__init__.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/configuration_conditional_detr.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/configuration_conditional_detr.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..45812e47ac4fc8ded30ac8e584b5e74972941ff6 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/configuration_conditional_detr.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/convert_conditional_detr_original_pytorch_checkpoint_to_pytorch.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/convert_conditional_detr_original_pytorch_checkpoint_to_pytorch.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..e4fbc1fbd43377f1f5bb6bd1985427bf4bf9acd5 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/convert_conditional_detr_original_pytorch_checkpoint_to_pytorch.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/feature_extraction_conditional_detr.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/feature_extraction_conditional_detr.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..897acf29e490b08e54e53df258d95cda4a75f9ec Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/feature_extraction_conditional_detr.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/image_processing_conditional_detr.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/image_processing_conditional_detr.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..842ecaed981eb9b4194514c6e06fb82e51bfbb5d Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/image_processing_conditional_detr.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/modeling_conditional_detr.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/modeling_conditional_detr.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..912789eb5887b8190730c68e5f45078867416299 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/__pycache__/modeling_conditional_detr.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/configuration_conditional_detr.py b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/configuration_conditional_detr.py new file mode 100644 index 0000000000000000000000000000000000000000..945e5edb32ad30a801d45e6de65edff1bfd2eae4 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/configuration_conditional_detr.py @@ -0,0 +1,273 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" Conditional DETR model configuration""" +from collections import OrderedDict +from typing import Mapping + +from packaging import version + +from ...configuration_utils import PretrainedConfig +from ...onnx import OnnxConfig +from ...utils import logging +from ..auto import CONFIG_MAPPING + + +logger = logging.get_logger(__name__) + + +from ..deprecated._archive_maps import CONDITIONAL_DETR_PRETRAINED_CONFIG_ARCHIVE_MAP # noqa: F401, E402 + + +class ConditionalDetrConfig(PretrainedConfig): + r""" + This is the configuration class to store the configuration of a [`ConditionalDetrModel`]. It is used to instantiate + a Conditional DETR model according to the specified arguments, defining the model architecture. Instantiating a + configuration with the defaults will yield a similar configuration to that of the Conditional DETR + [microsoft/conditional-detr-resnet-50](https://huggingface.co/microsoft/conditional-detr-resnet-50) architecture. + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + Args: + use_timm_backbone (`bool`, *optional*, defaults to `True`): + Whether or not to use the `timm` library for the backbone. If set to `False`, will use the [`AutoBackbone`] + API. + backbone_config (`PretrainedConfig` or `dict`, *optional*): + The configuration of the backbone model. Only used in case `use_timm_backbone` is set to `False` in which + case it will default to `ResNetConfig()`. + num_channels (`int`, *optional*, defaults to 3): + The number of input channels. + num_queries (`int`, *optional*, defaults to 100): + Number of object queries, i.e. detection slots. This is the maximal number of objects + [`ConditionalDetrModel`] can detect in a single image. For COCO, we recommend 100 queries. + d_model (`int`, *optional*, defaults to 256): + Dimension of the layers. + encoder_layers (`int`, *optional*, defaults to 6): + Number of encoder layers. + decoder_layers (`int`, *optional*, defaults to 6): + Number of decoder layers. + encoder_attention_heads (`int`, *optional*, defaults to 8): + Number of attention heads for each attention layer in the Transformer encoder. + decoder_attention_heads (`int`, *optional*, defaults to 8): + Number of attention heads for each attention layer in the Transformer decoder. + decoder_ffn_dim (`int`, *optional*, defaults to 2048): + Dimension of the "intermediate" (often named feed-forward) layer in decoder. + encoder_ffn_dim (`int`, *optional*, defaults to 2048): + Dimension of the "intermediate" (often named feed-forward) layer in decoder. + activation_function (`str` or `function`, *optional*, defaults to `"relu"`): + The non-linear activation function (function or string) in the encoder and pooler. If string, `"gelu"`, + `"relu"`, `"silu"` and `"gelu_new"` are supported. + dropout (`float`, *optional*, defaults to 0.1): + The dropout probability for all fully connected layers in the embeddings, encoder, and pooler. + attention_dropout (`float`, *optional*, defaults to 0.0): + The dropout ratio for the attention probabilities. + activation_dropout (`float`, *optional*, defaults to 0.0): + The dropout ratio for activations inside the fully connected layer. + init_std (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + init_xavier_std (`float`, *optional*, defaults to 1): + The scaling factor used for the Xavier initialization gain in the HM Attention map module. + encoder_layerdrop (`float`, *optional*, defaults to 0.0): + The LayerDrop probability for the encoder. See the [LayerDrop paper](see https://arxiv.org/abs/1909.11556) + for more details. + decoder_layerdrop (`float`, *optional*, defaults to 0.0): + The LayerDrop probability for the decoder. See the [LayerDrop paper](see https://arxiv.org/abs/1909.11556) + for more details. + auxiliary_loss (`bool`, *optional*, defaults to `False`): + Whether auxiliary decoding losses (loss at each decoder layer) are to be used. + position_embedding_type (`str`, *optional*, defaults to `"sine"`): + Type of position embeddings to be used on top of the image features. One of `"sine"` or `"learned"`. + backbone (`str`, *optional*, defaults to `"resnet50"`): + Name of backbone to use when `backbone_config` is `None`. If `use_pretrained_backbone` is `True`, this + will load the corresponding pretrained weights from the timm or transformers library. If `use_pretrained_backbone` + is `False`, this loads the backbone's config and uses that to initialize the backbone with random weights. + use_pretrained_backbone (`bool`, *optional*, defaults to `True`): + Whether to use pretrained weights for the backbone. + backbone_kwargs (`dict`, *optional*): + Keyword arguments to be passed to AutoBackbone when loading from a checkpoint + e.g. `{'out_indices': (0, 1, 2, 3)}`. Cannot be specified if `backbone_config` is set. + dilation (`bool`, *optional*, defaults to `False`): + Whether to replace stride with dilation in the last convolutional block (DC5). Only supported when + `use_timm_backbone` = `True`. + class_cost (`float`, *optional*, defaults to 1): + Relative weight of the classification error in the Hungarian matching cost. + bbox_cost (`float`, *optional*, defaults to 5): + Relative weight of the L1 error of the bounding box coordinates in the Hungarian matching cost. + giou_cost (`float`, *optional*, defaults to 2): + Relative weight of the generalized IoU loss of the bounding box in the Hungarian matching cost. + mask_loss_coefficient (`float`, *optional*, defaults to 1): + Relative weight of the Focal loss in the panoptic segmentation loss. + dice_loss_coefficient (`float`, *optional*, defaults to 1): + Relative weight of the DICE/F-1 loss in the panoptic segmentation loss. + bbox_loss_coefficient (`float`, *optional*, defaults to 5): + Relative weight of the L1 bounding box loss in the object detection loss. + giou_loss_coefficient (`float`, *optional*, defaults to 2): + Relative weight of the generalized IoU loss in the object detection loss. + eos_coefficient (`float`, *optional*, defaults to 0.1): + Relative classification weight of the 'no-object' class in the object detection loss. + focal_alpha (`float`, *optional*, defaults to 0.25): + Alpha parameter in the focal loss. + + Examples: + + ```python + >>> from transformers import ConditionalDetrConfig, ConditionalDetrModel + + >>> # Initializing a Conditional DETR microsoft/conditional-detr-resnet-50 style configuration + >>> configuration = ConditionalDetrConfig() + + >>> # Initializing a model (with random weights) from the microsoft/conditional-detr-resnet-50 style configuration + >>> model = ConditionalDetrModel(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + ```""" + + model_type = "conditional_detr" + keys_to_ignore_at_inference = ["past_key_values"] + attribute_map = { + "hidden_size": "d_model", + "num_attention_heads": "encoder_attention_heads", + } + + def __init__( + self, + use_timm_backbone=True, + backbone_config=None, + num_channels=3, + num_queries=300, + encoder_layers=6, + encoder_ffn_dim=2048, + encoder_attention_heads=8, + decoder_layers=6, + decoder_ffn_dim=2048, + decoder_attention_heads=8, + encoder_layerdrop=0.0, + decoder_layerdrop=0.0, + is_encoder_decoder=True, + activation_function="relu", + d_model=256, + dropout=0.1, + attention_dropout=0.0, + activation_dropout=0.0, + init_std=0.02, + init_xavier_std=1.0, + auxiliary_loss=False, + position_embedding_type="sine", + backbone="resnet50", + use_pretrained_backbone=True, + backbone_kwargs=None, + dilation=False, + class_cost=2, + bbox_cost=5, + giou_cost=2, + mask_loss_coefficient=1, + dice_loss_coefficient=1, + cls_loss_coefficient=2, + bbox_loss_coefficient=5, + giou_loss_coefficient=2, + focal_alpha=0.25, + **kwargs, + ): + if not use_timm_backbone and use_pretrained_backbone: + raise ValueError( + "Loading pretrained backbone weights from the transformers library is not supported yet. `use_timm_backbone` must be set to `True` when `use_pretrained_backbone=True`" + ) + + if backbone_config is not None and backbone is not None: + raise ValueError("You can't specify both `backbone` and `backbone_config`.") + + if backbone_config is not None and use_timm_backbone: + raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`.") + + if backbone_kwargs is not None and backbone_kwargs and backbone_config is not None: + raise ValueError("You can't specify both `backbone_kwargs` and `backbone_config`.") + + if not use_timm_backbone: + if backbone_config is None: + logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.") + backbone_config = CONFIG_MAPPING["resnet"](out_features=["stage4"]) + elif isinstance(backbone_config, dict): + backbone_model_type = backbone_config.get("model_type") + config_class = CONFIG_MAPPING[backbone_model_type] + backbone_config = config_class.from_dict(backbone_config) + + self.use_timm_backbone = use_timm_backbone + self.backbone_config = backbone_config + self.num_channels = num_channels + self.num_queries = num_queries + self.d_model = d_model + self.encoder_ffn_dim = encoder_ffn_dim + self.encoder_layers = encoder_layers + self.encoder_attention_heads = encoder_attention_heads + self.decoder_ffn_dim = decoder_ffn_dim + self.decoder_layers = decoder_layers + self.decoder_attention_heads = decoder_attention_heads + self.dropout = dropout + self.attention_dropout = attention_dropout + self.activation_dropout = activation_dropout + self.activation_function = activation_function + self.init_std = init_std + self.init_xavier_std = init_xavier_std + self.encoder_layerdrop = encoder_layerdrop + self.decoder_layerdrop = decoder_layerdrop + self.num_hidden_layers = encoder_layers + self.auxiliary_loss = auxiliary_loss + self.position_embedding_type = position_embedding_type + self.backbone = backbone + self.use_pretrained_backbone = use_pretrained_backbone + self.backbone_kwargs = backbone_kwargs + self.dilation = dilation + # Hungarian matcher + self.class_cost = class_cost + self.bbox_cost = bbox_cost + self.giou_cost = giou_cost + # Loss coefficients + self.mask_loss_coefficient = mask_loss_coefficient + self.dice_loss_coefficient = dice_loss_coefficient + self.cls_loss_coefficient = cls_loss_coefficient + self.bbox_loss_coefficient = bbox_loss_coefficient + self.giou_loss_coefficient = giou_loss_coefficient + self.focal_alpha = focal_alpha + super().__init__(is_encoder_decoder=is_encoder_decoder, **kwargs) + + @property + def num_attention_heads(self) -> int: + return self.encoder_attention_heads + + @property + def hidden_size(self) -> int: + return self.d_model + + +class ConditionalDetrOnnxConfig(OnnxConfig): + torch_onnx_minimum_version = version.parse("1.11") + + @property + def inputs(self) -> Mapping[str, Mapping[int, str]]: + return OrderedDict( + [ + ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), + ("pixel_mask", {0: "batch"}), + ] + ) + + @property + def atol_for_validation(self) -> float: + return 1e-5 + + @property + def default_onnx_opset(self) -> int: + return 12 diff --git a/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/image_processing_conditional_detr.py b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/image_processing_conditional_detr.py new file mode 100644 index 0000000000000000000000000000000000000000..e88bfc8fe230dffeb9d21e4057f15398e90c0222 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/conditional_detr/image_processing_conditional_detr.py @@ -0,0 +1,1777 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Image processor class for Conditional DETR.""" + +import io +import pathlib +from collections import defaultdict +from typing import Any, Callable, Dict, Iterable, List, Optional, Set, Tuple, Union + +import numpy as np + +from ...feature_extraction_utils import BatchFeature +from ...image_processing_utils import BaseImageProcessor, get_size_dict +from ...image_transforms import ( + PaddingMode, + center_to_corners_format, + corners_to_center_format, + id_to_rgb, + pad, + rescale, + resize, + rgb_to_id, + to_channel_dimension_format, +) +from ...image_utils import ( + IMAGENET_DEFAULT_MEAN, + IMAGENET_DEFAULT_STD, + AnnotationFormat, + AnnotationType, + ChannelDimension, + ImageInput, + PILImageResampling, + get_image_size, + infer_channel_dimension_format, + is_scaled_image, + make_list_of_images, + to_numpy_array, + valid_images, + validate_annotations, + validate_kwargs, + validate_preprocess_arguments, +) +from ...utils import ( + TensorType, + is_flax_available, + is_jax_tensor, + is_scipy_available, + is_tf_available, + is_tf_tensor, + is_torch_available, + is_torch_tensor, + is_vision_available, + logging, +) + + +if is_torch_available(): + import torch + from torch import nn + + +if is_vision_available(): + import PIL + + +if is_scipy_available(): + import scipy.special + import scipy.stats + + +logger = logging.get_logger(__name__) # pylint: disable=invalid-name + + +SUPPORTED_ANNOTATION_FORMATS = (AnnotationFormat.COCO_DETECTION, AnnotationFormat.COCO_PANOPTIC) + + +# Copied from transformers.models.detr.image_processing_detr.get_size_with_aspect_ratio +def get_size_with_aspect_ratio(image_size, size, max_size=None) -> Tuple[int, int]: + """ + Computes the output image size given the input image size and the desired output size. + + Args: + image_size (`Tuple[int, int]`): + The input image size. + size (`int`): + The desired output size. + max_size (`int`, *optional*): + The maximum allowed output size. + """ + height, width = image_size + if max_size is not None: + min_original_size = float(min((height, width))) + max_original_size = float(max((height, width))) + if max_original_size / min_original_size * size > max_size: + size = int(round(max_size * min_original_size / max_original_size)) + + if (height <= width and height == size) or (width <= height and width == size): + return height, width + + if width < height: + ow = size + oh = int(size * height / width) + else: + oh = size + ow = int(size * width / height) + return (oh, ow) + + +# Copied from transformers.models.detr.image_processing_detr.get_resize_output_image_size +def get_resize_output_image_size( + input_image: np.ndarray, + size: Union[int, Tuple[int, int], List[int]], + max_size: Optional[int] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, +) -> Tuple[int, int]: + """ + Computes the output image size given the input image size and the desired output size. If the desired output size + is a tuple or list, the output image size is returned as is. If the desired output size is an integer, the output + image size is computed by keeping the aspect ratio of the input image size. + + Args: + input_image (`np.ndarray`): + The image to resize. + size (`int` or `Tuple[int, int]` or `List[int]`): + The desired output size. + max_size (`int`, *optional*): + The maximum allowed output size. + input_data_format (`ChannelDimension` or `str`, *optional*): + The channel dimension format of the input image. If not provided, it will be inferred from the input image. + """ + image_size = get_image_size(input_image, input_data_format) + if isinstance(size, (list, tuple)): + return size + + return get_size_with_aspect_ratio(image_size, size, max_size) + + +# Copied from transformers.models.detr.image_processing_detr.get_numpy_to_framework_fn +def get_numpy_to_framework_fn(arr) -> Callable: + """ + Returns a function that converts a numpy array to the framework of the input array. + + Args: + arr (`np.ndarray`): The array to convert. + """ + if isinstance(arr, np.ndarray): + return np.array + if is_tf_available() and is_tf_tensor(arr): + import tensorflow as tf + + return tf.convert_to_tensor + if is_torch_available() and is_torch_tensor(arr): + import torch + + return torch.tensor + if is_flax_available() and is_jax_tensor(arr): + import jax.numpy as jnp + + return jnp.array + raise ValueError(f"Cannot convert arrays of type {type(arr)}") + + +# Copied from transformers.models.detr.image_processing_detr.safe_squeeze +def safe_squeeze(arr: np.ndarray, axis: Optional[int] = None) -> np.ndarray: + """ + Squeezes an array, but only if the axis specified has dim 1. + """ + if axis is None: + return arr.squeeze() + + try: + return arr.squeeze(axis=axis) + except ValueError: + return arr + + +# Copied from transformers.models.detr.image_processing_detr.normalize_annotation +def normalize_annotation(annotation: Dict, image_size: Tuple[int, int]) -> Dict: + image_height, image_width = image_size + norm_annotation = {} + for key, value in annotation.items(): + if key == "boxes": + boxes = value + boxes = corners_to_center_format(boxes) + boxes /= np.asarray([image_width, image_height, image_width, image_height], dtype=np.float32) + norm_annotation[key] = boxes + else: + norm_annotation[key] = value + return norm_annotation + + +# Copied from transformers.models.detr.image_processing_detr.max_across_indices +def max_across_indices(values: Iterable[Any]) -> List[Any]: + """ + Return the maximum value across all indices of an iterable of values. + """ + return [max(values_i) for values_i in zip(*values)] + + +# Copied from transformers.models.detr.image_processing_detr.get_max_height_width +def get_max_height_width( + images: List[np.ndarray], input_data_format: Optional[Union[str, ChannelDimension]] = None +) -> List[int]: + """ + Get the maximum height and width across all images in a batch. + """ + if input_data_format is None: + input_data_format = infer_channel_dimension_format(images[0]) + + if input_data_format == ChannelDimension.FIRST: + _, max_height, max_width = max_across_indices([img.shape for img in images]) + elif input_data_format == ChannelDimension.LAST: + max_height, max_width, _ = max_across_indices([img.shape for img in images]) + else: + raise ValueError(f"Invalid channel dimension format: {input_data_format}") + return (max_height, max_width) + + +# Copied from transformers.models.detr.image_processing_detr.make_pixel_mask +def make_pixel_mask( + image: np.ndarray, output_size: Tuple[int, int], input_data_format: Optional[Union[str, ChannelDimension]] = None +) -> np.ndarray: + """ + Make a pixel mask for the image, where 1 indicates a valid pixel and 0 indicates padding. + + Args: + image (`np.ndarray`): + Image to make the pixel mask for. + output_size (`Tuple[int, int]`): + Output size of the mask. + """ + input_height, input_width = get_image_size(image, channel_dim=input_data_format) + mask = np.zeros(output_size, dtype=np.int64) + mask[:input_height, :input_width] = 1 + return mask + + +# Copied from transformers.models.detr.image_processing_detr.convert_coco_poly_to_mask +def convert_coco_poly_to_mask(segmentations, height: int, width: int) -> np.ndarray: + """ + Convert a COCO polygon annotation to a mask. + + Args: + segmentations (`List[List[float]]`): + List of polygons, each polygon represented by a list of x-y coordinates. + height (`int`): + Height of the mask. + width (`int`): + Width of the mask. + """ + try: + from pycocotools import mask as coco_mask + except ImportError: + raise ImportError("Pycocotools is not installed in your environment.") + + masks = [] + for polygons in segmentations: + rles = coco_mask.frPyObjects(polygons, height, width) + mask = coco_mask.decode(rles) + if len(mask.shape) < 3: + mask = mask[..., None] + mask = np.asarray(mask, dtype=np.uint8) + mask = np.any(mask, axis=2) + masks.append(mask) + if masks: + masks = np.stack(masks, axis=0) + else: + masks = np.zeros((0, height, width), dtype=np.uint8) + + return masks + + +# Copied from transformers.models.detr.image_processing_detr.prepare_coco_detection_annotation with DETR->ConditionalDetr +def prepare_coco_detection_annotation( + image, + target, + return_segmentation_masks: bool = False, + input_data_format: Optional[Union[ChannelDimension, str]] = None, +): + """ + Convert the target in COCO format into the format expected by ConditionalDetr. + """ + image_height, image_width = get_image_size(image, channel_dim=input_data_format) + + image_id = target["image_id"] + image_id = np.asarray([image_id], dtype=np.int64) + + # Get all COCO annotations for the given image. + annotations = target["annotations"] + annotations = [obj for obj in annotations if "iscrowd" not in obj or obj["iscrowd"] == 0] + + classes = [obj["category_id"] for obj in annotations] + classes = np.asarray(classes, dtype=np.int64) + + # for conversion to coco api + area = np.asarray([obj["area"] for obj in annotations], dtype=np.float32) + iscrowd = np.asarray([obj["iscrowd"] if "iscrowd" in obj else 0 for obj in annotations], dtype=np.int64) + + boxes = [obj["bbox"] for obj in annotations] + # guard against no boxes via resizing + boxes = np.asarray(boxes, dtype=np.float32).reshape(-1, 4) + boxes[:, 2:] += boxes[:, :2] + boxes[:, 0::2] = boxes[:, 0::2].clip(min=0, max=image_width) + boxes[:, 1::2] = boxes[:, 1::2].clip(min=0, max=image_height) + + keep = (boxes[:, 3] > boxes[:, 1]) & (boxes[:, 2] > boxes[:, 0]) + + new_target = {} + new_target["image_id"] = image_id + new_target["class_labels"] = classes[keep] + new_target["boxes"] = boxes[keep] + new_target["area"] = area[keep] + new_target["iscrowd"] = iscrowd[keep] + new_target["orig_size"] = np.asarray([int(image_height), int(image_width)], dtype=np.int64) + + if annotations and "keypoints" in annotations[0]: + keypoints = [obj["keypoints"] for obj in annotations] + # Converting the filtered keypoints list to a numpy array + keypoints = np.asarray(keypoints, dtype=np.float32) + # Apply the keep mask here to filter the relevant annotations + keypoints = keypoints[keep] + num_keypoints = keypoints.shape[0] + keypoints = keypoints.reshape((-1, 3)) if num_keypoints else keypoints + new_target["keypoints"] = keypoints + + if return_segmentation_masks: + segmentation_masks = [obj["segmentation"] for obj in annotations] + masks = convert_coco_poly_to_mask(segmentation_masks, image_height, image_width) + new_target["masks"] = masks[keep] + + return new_target + + +# Copied from transformers.models.detr.image_processing_detr.masks_to_boxes +def masks_to_boxes(masks: np.ndarray) -> np.ndarray: + """ + Compute the bounding boxes around the provided panoptic segmentation masks. + + Args: + masks: masks in format `[number_masks, height, width]` where N is the number of masks + + Returns: + boxes: bounding boxes in format `[number_masks, 4]` in xyxy format + """ + if masks.size == 0: + return np.zeros((0, 4)) + + h, w = masks.shape[-2:] + y = np.arange(0, h, dtype=np.float32) + x = np.arange(0, w, dtype=np.float32) + # see https://github.com/pytorch/pytorch/issues/50276 + y, x = np.meshgrid(y, x, indexing="ij") + + x_mask = masks * np.expand_dims(x, axis=0) + x_max = x_mask.reshape(x_mask.shape[0], -1).max(-1) + x = np.ma.array(x_mask, mask=~(np.array(masks, dtype=bool))) + x_min = x.filled(fill_value=1e8) + x_min = x_min.reshape(x_min.shape[0], -1).min(-1) + + y_mask = masks * np.expand_dims(y, axis=0) + y_max = y_mask.reshape(x_mask.shape[0], -1).max(-1) + y = np.ma.array(y_mask, mask=~(np.array(masks, dtype=bool))) + y_min = y.filled(fill_value=1e8) + y_min = y_min.reshape(y_min.shape[0], -1).min(-1) + + return np.stack([x_min, y_min, x_max, y_max], 1) + + +# Copied from transformers.models.detr.image_processing_detr.prepare_coco_panoptic_annotation with DETR->ConditionalDetr +def prepare_coco_panoptic_annotation( + image: np.ndarray, + target: Dict, + masks_path: Union[str, pathlib.Path], + return_masks: bool = True, + input_data_format: Union[ChannelDimension, str] = None, +) -> Dict: + """ + Prepare a coco panoptic annotation for ConditionalDetr. + """ + image_height, image_width = get_image_size(image, channel_dim=input_data_format) + annotation_path = pathlib.Path(masks_path) / target["file_name"] + + new_target = {} + new_target["image_id"] = np.asarray([target["image_id"] if "image_id" in target else target["id"]], dtype=np.int64) + new_target["size"] = np.asarray([image_height, image_width], dtype=np.int64) + new_target["orig_size"] = np.asarray([image_height, image_width], dtype=np.int64) + + if "segments_info" in target: + masks = np.asarray(PIL.Image.open(annotation_path), dtype=np.uint32) + masks = rgb_to_id(masks) + + ids = np.array([segment_info["id"] for segment_info in target["segments_info"]]) + masks = masks == ids[:, None, None] + masks = masks.astype(np.uint8) + if return_masks: + new_target["masks"] = masks + new_target["boxes"] = masks_to_boxes(masks) + new_target["class_labels"] = np.array( + [segment_info["category_id"] for segment_info in target["segments_info"]], dtype=np.int64 + ) + new_target["iscrowd"] = np.asarray( + [segment_info["iscrowd"] for segment_info in target["segments_info"]], dtype=np.int64 + ) + new_target["area"] = np.asarray( + [segment_info["area"] for segment_info in target["segments_info"]], dtype=np.float32 + ) + + return new_target + + +# Copied from transformers.models.detr.image_processing_detr.get_segmentation_image +def get_segmentation_image( + masks: np.ndarray, input_size: Tuple, target_size: Tuple, stuff_equiv_classes, deduplicate=False +): + h, w = input_size + final_h, final_w = target_size + + m_id = scipy.special.softmax(masks.transpose(0, 1), -1) + + if m_id.shape[-1] == 0: + # We didn't detect any mask :( + m_id = np.zeros((h, w), dtype=np.int64) + else: + m_id = m_id.argmax(-1).reshape(h, w) + + if deduplicate: + # Merge the masks corresponding to the same stuff class + for equiv in stuff_equiv_classes.values(): + for eq_id in equiv: + m_id[m_id == eq_id] = equiv[0] + + seg_img = id_to_rgb(m_id) + seg_img = resize(seg_img, (final_w, final_h), resample=PILImageResampling.NEAREST) + return seg_img + + +# Copied from transformers.models.detr.image_processing_detr.get_mask_area +def get_mask_area(seg_img: np.ndarray, target_size: Tuple[int, int], n_classes: int) -> np.ndarray: + final_h, final_w = target_size + np_seg_img = seg_img.astype(np.uint8) + np_seg_img = np_seg_img.reshape(final_h, final_w, 3) + m_id = rgb_to_id(np_seg_img) + area = [(m_id == i).sum() for i in range(n_classes)] + return area + + +# Copied from transformers.models.detr.image_processing_detr.score_labels_from_class_probabilities +def score_labels_from_class_probabilities(logits: np.ndarray) -> Tuple[np.ndarray, np.ndarray]: + probs = scipy.special.softmax(logits, axis=-1) + labels = probs.argmax(-1, keepdims=True) + scores = np.take_along_axis(probs, labels, axis=-1) + scores, labels = scores.squeeze(-1), labels.squeeze(-1) + return scores, labels + + +# Copied from transformers.models.detr.image_processing_detr.post_process_panoptic_sample with DetrForSegmentation->ConditionalDetrForSegmentation +def post_process_panoptic_sample( + out_logits: np.ndarray, + masks: np.ndarray, + boxes: np.ndarray, + processed_size: Tuple[int, int], + target_size: Tuple[int, int], + is_thing_map: Dict, + threshold=0.85, +) -> Dict: + """ + Converts the output of [`ConditionalDetrForSegmentation`] into panoptic segmentation predictions for a single sample. + + Args: + out_logits (`torch.Tensor`): + The logits for this sample. + masks (`torch.Tensor`): + The predicted segmentation masks for this sample. + boxes (`torch.Tensor`): + The prediced bounding boxes for this sample. The boxes are in the normalized format `(center_x, center_y, + width, height)` and values between `[0, 1]`, relative to the size the image (disregarding padding). + processed_size (`Tuple[int, int]`): + The processed size of the image `(height, width)`, as returned by the preprocessing step i.e. the size + after data augmentation but before batching. + target_size (`Tuple[int, int]`): + The target size of the image, `(height, width)` corresponding to the requested final size of the + prediction. + is_thing_map (`Dict`): + A dictionary mapping class indices to a boolean value indicating whether the class is a thing or not. + threshold (`float`, *optional*, defaults to 0.85): + The threshold used to binarize the segmentation masks. + """ + # we filter empty queries and detection below threshold + scores, labels = score_labels_from_class_probabilities(out_logits) + keep = (labels != out_logits.shape[-1] - 1) & (scores > threshold) + + cur_scores = scores[keep] + cur_classes = labels[keep] + cur_boxes = center_to_corners_format(boxes[keep]) + + if len(cur_boxes) != len(cur_classes): + raise ValueError("Not as many boxes as there are classes") + + cur_masks = masks[keep] + cur_masks = resize(cur_masks[:, None], processed_size, resample=PILImageResampling.BILINEAR) + cur_masks = safe_squeeze(cur_masks, 1) + b, h, w = cur_masks.shape + + # It may be that we have several predicted masks for the same stuff class. + # In the following, we track the list of masks ids for each stuff class (they are merged later on) + cur_masks = cur_masks.reshape(b, -1) + stuff_equiv_classes = defaultdict(list) + for k, label in enumerate(cur_classes): + if not is_thing_map[label]: + stuff_equiv_classes[label].append(k) + + seg_img = get_segmentation_image(cur_masks, processed_size, target_size, stuff_equiv_classes, deduplicate=True) + area = get_mask_area(cur_masks, processed_size, n_classes=len(cur_scores)) + + # We filter out any mask that is too small + if cur_classes.size() > 0: + # We know filter empty masks as long as we find some + filtered_small = np.array([a <= 4 for a in area], dtype=bool) + while filtered_small.any(): + cur_masks = cur_masks[~filtered_small] + cur_scores = cur_scores[~filtered_small] + cur_classes = cur_classes[~filtered_small] + seg_img = get_segmentation_image(cur_masks, (h, w), target_size, stuff_equiv_classes, deduplicate=True) + area = get_mask_area(seg_img, target_size, n_classes=len(cur_scores)) + filtered_small = np.array([a <= 4 for a in area], dtype=bool) + else: + cur_classes = np.ones((1, 1), dtype=np.int64) + + segments_info = [ + {"id": i, "isthing": is_thing_map[cat], "category_id": int(cat), "area": a} + for i, (cat, a) in enumerate(zip(cur_classes, area)) + ] + del cur_classes + + with io.BytesIO() as out: + PIL.Image.fromarray(seg_img).save(out, format="PNG") + predictions = {"png_string": out.getvalue(), "segments_info": segments_info} + + return predictions + + +# Copied from transformers.models.detr.image_processing_detr.resize_annotation +def resize_annotation( + annotation: Dict[str, Any], + orig_size: Tuple[int, int], + target_size: Tuple[int, int], + threshold: float = 0.5, + resample: PILImageResampling = PILImageResampling.NEAREST, +): + """ + Resizes an annotation to a target size. + + Args: + annotation (`Dict[str, Any]`): + The annotation dictionary. + orig_size (`Tuple[int, int]`): + The original size of the input image. + target_size (`Tuple[int, int]`): + The target size of the image, as returned by the preprocessing `resize` step. + threshold (`float`, *optional*, defaults to 0.5): + The threshold used to binarize the segmentation masks. + resample (`PILImageResampling`, defaults to `PILImageResampling.NEAREST`): + The resampling filter to use when resizing the masks. + """ + ratios = tuple(float(s) / float(s_orig) for s, s_orig in zip(target_size, orig_size)) + ratio_height, ratio_width = ratios + + new_annotation = {} + new_annotation["size"] = target_size + + for key, value in annotation.items(): + if key == "boxes": + boxes = value + scaled_boxes = boxes * np.asarray([ratio_width, ratio_height, ratio_width, ratio_height], dtype=np.float32) + new_annotation["boxes"] = scaled_boxes + elif key == "area": + area = value + scaled_area = area * (ratio_width * ratio_height) + new_annotation["area"] = scaled_area + elif key == "masks": + masks = value[:, None] + masks = np.array([resize(mask, target_size, resample=resample) for mask in masks]) + masks = masks.astype(np.float32) + masks = masks[:, 0] > threshold + new_annotation["masks"] = masks + elif key == "size": + new_annotation["size"] = target_size + else: + new_annotation[key] = value + + return new_annotation + + +# Copied from transformers.models.detr.image_processing_detr.binary_mask_to_rle +def binary_mask_to_rle(mask): + """ + Converts given binary mask of shape `(height, width)` to the run-length encoding (RLE) format. + + Args: + mask (`torch.Tensor` or `numpy.array`): + A binary mask tensor of shape `(height, width)` where 0 denotes background and 1 denotes the target + segment_id or class_id. + Returns: + `List`: Run-length encoded list of the binary mask. Refer to COCO API for more information about the RLE + format. + """ + if is_torch_tensor(mask): + mask = mask.numpy() + + pixels = mask.flatten() + pixels = np.concatenate([[0], pixels, [0]]) + runs = np.where(pixels[1:] != pixels[:-1])[0] + 1 + runs[1::2] -= runs[::2] + return list(runs) + + +# Copied from transformers.models.detr.image_processing_detr.convert_segmentation_to_rle +def convert_segmentation_to_rle(segmentation): + """ + Converts given segmentation map of shape `(height, width)` to the run-length encoding (RLE) format. + + Args: + segmentation (`torch.Tensor` or `numpy.array`): + A segmentation map of shape `(height, width)` where each value denotes a segment or class id. + Returns: + `List[List]`: A list of lists, where each list is the run-length encoding of a segment / class id. + """ + segment_ids = torch.unique(segmentation) + + run_length_encodings = [] + for idx in segment_ids: + mask = torch.where(segmentation == idx, 1, 0) + rle = binary_mask_to_rle(mask) + run_length_encodings.append(rle) + + return run_length_encodings + + +# Copied from transformers.models.detr.image_processing_detr.remove_low_and_no_objects +def remove_low_and_no_objects(masks, scores, labels, object_mask_threshold, num_labels): + """ + Binarize the given masks using `object_mask_threshold`, it returns the associated values of `masks`, `scores` and + `labels`. + + Args: + masks (`torch.Tensor`): + A tensor of shape `(num_queries, height, width)`. + scores (`torch.Tensor`): + A tensor of shape `(num_queries)`. + labels (`torch.Tensor`): + A tensor of shape `(num_queries)`. + object_mask_threshold (`float`): + A number between 0 and 1 used to binarize the masks. + Raises: + `ValueError`: Raised when the first dimension doesn't match in all input tensors. + Returns: + `Tuple[`torch.Tensor`, `torch.Tensor`, `torch.Tensor`]`: The `masks`, `scores` and `labels` without the region + < `object_mask_threshold`. + """ + if not (masks.shape[0] == scores.shape[0] == labels.shape[0]): + raise ValueError("mask, scores and labels must have the same shape!") + + to_keep = labels.ne(num_labels) & (scores > object_mask_threshold) + + return masks[to_keep], scores[to_keep], labels[to_keep] + + +# Copied from transformers.models.detr.image_processing_detr.check_segment_validity +def check_segment_validity(mask_labels, mask_probs, k, mask_threshold=0.5, overlap_mask_area_threshold=0.8): + # Get the mask associated with the k class + mask_k = mask_labels == k + mask_k_area = mask_k.sum() + + # Compute the area of all the stuff in query k + original_area = (mask_probs[k] >= mask_threshold).sum() + mask_exists = mask_k_area > 0 and original_area > 0 + + # Eliminate disconnected tiny segments + if mask_exists: + area_ratio = mask_k_area / original_area + if not area_ratio.item() > overlap_mask_area_threshold: + mask_exists = False + + return mask_exists, mask_k + + +# Copied from transformers.models.detr.image_processing_detr.compute_segments +def compute_segments( + mask_probs, + pred_scores, + pred_labels, + mask_threshold: float = 0.5, + overlap_mask_area_threshold: float = 0.8, + label_ids_to_fuse: Optional[Set[int]] = None, + target_size: Tuple[int, int] = None, +): + height = mask_probs.shape[1] if target_size is None else target_size[0] + width = mask_probs.shape[2] if target_size is None else target_size[1] + + segmentation = torch.zeros((height, width), dtype=torch.int32, device=mask_probs.device) + segments: List[Dict] = [] + + if target_size is not None: + mask_probs = nn.functional.interpolate( + mask_probs.unsqueeze(0), size=target_size, mode="bilinear", align_corners=False + )[0] + + current_segment_id = 0 + + # Weigh each mask by its prediction score + mask_probs *= pred_scores.view(-1, 1, 1) + mask_labels = mask_probs.argmax(0) # [height, width] + + # Keep track of instances of each class + stuff_memory_list: Dict[str, int] = {} + for k in range(pred_labels.shape[0]): + pred_class = pred_labels[k].item() + should_fuse = pred_class in label_ids_to_fuse + + # Check if mask exists and large enough to be a segment + mask_exists, mask_k = check_segment_validity( + mask_labels, mask_probs, k, mask_threshold, overlap_mask_area_threshold + ) + + if mask_exists: + if pred_class in stuff_memory_list: + current_segment_id = stuff_memory_list[pred_class] + else: + current_segment_id += 1 + + # Add current object segment to final segmentation map + segmentation[mask_k] = current_segment_id + segment_score = round(pred_scores[k].item(), 6) + segments.append( + { + "id": current_segment_id, + "label_id": pred_class, + "was_fused": should_fuse, + "score": segment_score, + } + ) + if should_fuse: + stuff_memory_list[pred_class] = current_segment_id + + return segmentation, segments + + +class ConditionalDetrImageProcessor(BaseImageProcessor): + r""" + Constructs a Conditional Detr image processor. + + Args: + format (`str`, *optional*, defaults to `"coco_detection"`): + Data format of the annotations. One of "coco_detection" or "coco_panoptic". + do_resize (`bool`, *optional*, defaults to `True`): + Controls whether to resize the image's (height, width) dimensions to the specified `size`. Can be + overridden by the `do_resize` parameter in the `preprocess` method. + size (`Dict[str, int]` *optional*, defaults to `{"shortest_edge": 800, "longest_edge": 1333}`): + Size of the image's (height, width) dimensions after resizing. Can be overridden by the `size` parameter in + the `preprocess` method. + resample (`PILImageResampling`, *optional*, defaults to `PILImageResampling.BILINEAR`): + Resampling filter to use if resizing the image. + do_rescale (`bool`, *optional*, defaults to `True`): + Controls whether to rescale the image by the specified scale `rescale_factor`. Can be overridden by the + `do_rescale` parameter in the `preprocess` method. + rescale_factor (`int` or `float`, *optional*, defaults to `1/255`): + Scale factor to use if rescaling the image. Can be overridden by the `rescale_factor` parameter in the + `preprocess` method. + do_normalize: + Controls whether to normalize the image. Can be overridden by the `do_normalize` parameter in the + `preprocess` method. + image_mean (`float` or `List[float]`, *optional*, defaults to `IMAGENET_DEFAULT_MEAN`): + Mean values to use when normalizing the image. Can be a single value or a list of values, one for each + channel. Can be overridden by the `image_mean` parameter in the `preprocess` method. + image_std (`float` or `List[float]`, *optional*, defaults to `IMAGENET_DEFAULT_STD`): + Standard deviation values to use when normalizing the image. Can be a single value or a list of values, one + for each channel. Can be overridden by the `image_std` parameter in the `preprocess` method. + do_convert_annotations (`bool`, *optional*, defaults to `True`): + Controls whether to convert the annotations to the format expected by the DETR model. Converts the + bounding boxes to the format `(center_x, center_y, width, height)` and in the range `[0, 1]`. + Can be overridden by the `do_convert_annotations` parameter in the `preprocess` method. + do_pad (`bool`, *optional*, defaults to `True`): + Controls whether to pad the image. Can be overridden by the `do_pad` parameter in the `preprocess` + method. If `True` will pad the images in the batch to the largest height and width in the batch. + Padding will be applied to the bottom and right of the image with zeros. + """ + + model_input_names = ["pixel_values", "pixel_mask"] + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.__init__ + def __init__( + self, + format: Union[str, AnnotationFormat] = AnnotationFormat.COCO_DETECTION, + do_resize: bool = True, + size: Dict[str, int] = None, + resample: PILImageResampling = PILImageResampling.BILINEAR, + do_rescale: bool = True, + rescale_factor: Union[int, float] = 1 / 255, + do_normalize: bool = True, + image_mean: Union[float, List[float]] = None, + image_std: Union[float, List[float]] = None, + do_convert_annotations: Optional[bool] = None, + do_pad: bool = True, + **kwargs, + ) -> None: + if "pad_and_return_pixel_mask" in kwargs: + do_pad = kwargs.pop("pad_and_return_pixel_mask") + + if "max_size" in kwargs: + logger.warning_once( + "The `max_size` parameter is deprecated and will be removed in v4.26. " + "Please specify in `size['longest_edge'] instead`.", + ) + max_size = kwargs.pop("max_size") + else: + max_size = None if size is None else 1333 + + size = size if size is not None else {"shortest_edge": 800, "longest_edge": 1333} + size = get_size_dict(size, max_size=max_size, default_to_square=False) + + # Backwards compatibility + if do_convert_annotations is None: + do_convert_annotations = do_normalize + + super().__init__(**kwargs) + self.format = format + self.do_resize = do_resize + self.size = size + self.resample = resample + self.do_rescale = do_rescale + self.rescale_factor = rescale_factor + self.do_normalize = do_normalize + self.do_convert_annotations = do_convert_annotations + self.image_mean = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN + self.image_std = image_std if image_std is not None else IMAGENET_DEFAULT_STD + self.do_pad = do_pad + self._valid_processor_keys = [ + "images", + "annotations", + "return_segmentation_masks", + "masks_path", + "do_resize", + "size", + "resample", + "do_rescale", + "rescale_factor", + "do_normalize", + "do_convert_annotations", + "image_mean", + "image_std", + "do_pad", + "format", + "return_tensors", + "data_format", + "input_data_format", + ] + + @classmethod + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.from_dict with Detr->ConditionalDetr + def from_dict(cls, image_processor_dict: Dict[str, Any], **kwargs): + """ + Overrides the `from_dict` method from the base class to make sure parameters are updated if image processor is + created using from_dict and kwargs e.g. `ConditionalDetrImageProcessor.from_pretrained(checkpoint, size=600, + max_size=800)` + """ + image_processor_dict = image_processor_dict.copy() + if "max_size" in kwargs: + image_processor_dict["max_size"] = kwargs.pop("max_size") + if "pad_and_return_pixel_mask" in kwargs: + image_processor_dict["pad_and_return_pixel_mask"] = kwargs.pop("pad_and_return_pixel_mask") + return super().from_dict(image_processor_dict, **kwargs) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.prepare_annotation with DETR->ConditionalDetr + def prepare_annotation( + self, + image: np.ndarray, + target: Dict, + format: Optional[AnnotationFormat] = None, + return_segmentation_masks: bool = None, + masks_path: Optional[Union[str, pathlib.Path]] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + ) -> Dict: + """ + Prepare an annotation for feeding into ConditionalDetr model. + """ + format = format if format is not None else self.format + + if format == AnnotationFormat.COCO_DETECTION: + return_segmentation_masks = False if return_segmentation_masks is None else return_segmentation_masks + target = prepare_coco_detection_annotation( + image, target, return_segmentation_masks, input_data_format=input_data_format + ) + elif format == AnnotationFormat.COCO_PANOPTIC: + return_segmentation_masks = True if return_segmentation_masks is None else return_segmentation_masks + target = prepare_coco_panoptic_annotation( + image, + target, + masks_path=masks_path, + return_masks=return_segmentation_masks, + input_data_format=input_data_format, + ) + else: + raise ValueError(f"Format {format} is not supported.") + return target + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.prepare + def prepare(self, image, target, return_segmentation_masks=None, masks_path=None): + logger.warning_once( + "The `prepare` method is deprecated and will be removed in a v4.33. " + "Please use `prepare_annotation` instead. Note: the `prepare_annotation` method " + "does not return the image anymore.", + ) + target = self.prepare_annotation(image, target, return_segmentation_masks, masks_path, self.format) + return image, target + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.convert_coco_poly_to_mask + def convert_coco_poly_to_mask(self, *args, **kwargs): + logger.warning_once("The `convert_coco_poly_to_mask` method is deprecated and will be removed in v4.33. ") + return convert_coco_poly_to_mask(*args, **kwargs) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.prepare_coco_detection with DETR->ConditionalDetr + def prepare_coco_detection(self, *args, **kwargs): + logger.warning_once("The `prepare_coco_detection` method is deprecated and will be removed in v4.33. ") + return prepare_coco_detection_annotation(*args, **kwargs) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.prepare_coco_panoptic + def prepare_coco_panoptic(self, *args, **kwargs): + logger.warning_once("The `prepare_coco_panoptic` method is deprecated and will be removed in v4.33. ") + return prepare_coco_panoptic_annotation(*args, **kwargs) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.resize + def resize( + self, + image: np.ndarray, + size: Dict[str, int], + resample: PILImageResampling = PILImageResampling.BILINEAR, + data_format: Optional[ChannelDimension] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + **kwargs, + ) -> np.ndarray: + """ + Resize the image to the given size. Size can be `min_size` (scalar) or `(height, width)` tuple. If size is an + int, smaller edge of the image will be matched to this number. + + Args: + image (`np.ndarray`): + Image to resize. + size (`Dict[str, int]`): + Dictionary containing the size to resize to. Can contain the keys `shortest_edge` and `longest_edge` or + `height` and `width`. + resample (`PILImageResampling`, *optional*, defaults to `PILImageResampling.BILINEAR`): + Resampling filter to use if resizing the image. + data_format (`str` or `ChannelDimension`, *optional*): + The channel dimension format for the output image. If unset, the channel dimension format of the input + image is used. + input_data_format (`ChannelDimension` or `str`, *optional*): + The channel dimension format of the input image. If not provided, it will be inferred. + """ + if "max_size" in kwargs: + logger.warning_once( + "The `max_size` parameter is deprecated and will be removed in v4.26. " + "Please specify in `size['longest_edge'] instead`.", + ) + max_size = kwargs.pop("max_size") + else: + max_size = None + size = get_size_dict(size, max_size=max_size, default_to_square=False) + if "shortest_edge" in size and "longest_edge" in size: + size = get_resize_output_image_size( + image, size["shortest_edge"], size["longest_edge"], input_data_format=input_data_format + ) + elif "height" in size and "width" in size: + size = (size["height"], size["width"]) + else: + raise ValueError( + "Size must contain 'height' and 'width' keys or 'shortest_edge' and 'longest_edge' keys. Got" + f" {size.keys()}." + ) + image = resize( + image, size=size, resample=resample, data_format=data_format, input_data_format=input_data_format, **kwargs + ) + return image + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.resize_annotation + def resize_annotation( + self, + annotation, + orig_size, + size, + resample: PILImageResampling = PILImageResampling.NEAREST, + ) -> Dict: + """ + Resize the annotation to match the resized image. If size is an int, smaller edge of the mask will be matched + to this number. + """ + return resize_annotation(annotation, orig_size=orig_size, target_size=size, resample=resample) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.rescale + def rescale( + self, + image: np.ndarray, + rescale_factor: float, + data_format: Optional[Union[str, ChannelDimension]] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + ) -> np.ndarray: + """ + Rescale the image by the given factor. image = image * rescale_factor. + + Args: + image (`np.ndarray`): + Image to rescale. + rescale_factor (`float`): + The value to use for rescaling. + data_format (`str` or `ChannelDimension`, *optional*): + The channel dimension format for the output image. If unset, the channel dimension format of the input + image is used. Can be one of: + - `"channels_first"` or `ChannelDimension.FIRST`: image in (num_channels, height, width) format. + - `"channels_last"` or `ChannelDimension.LAST`: image in (height, width, num_channels) format. + input_data_format (`str` or `ChannelDimension`, *optional*): + The channel dimension format for the input image. If unset, is inferred from the input image. Can be + one of: + - `"channels_first"` or `ChannelDimension.FIRST`: image in (num_channels, height, width) format. + - `"channels_last"` or `ChannelDimension.LAST`: image in (height, width, num_channels) format. + """ + return rescale(image, rescale_factor, data_format=data_format, input_data_format=input_data_format) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.normalize_annotation + def normalize_annotation(self, annotation: Dict, image_size: Tuple[int, int]) -> Dict: + """ + Normalize the boxes in the annotation from `[top_left_x, top_left_y, bottom_right_x, bottom_right_y]` to + `[center_x, center_y, width, height]` format and from absolute to relative pixel values. + """ + return normalize_annotation(annotation, image_size=image_size) + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor._update_annotation_for_padded_image + def _update_annotation_for_padded_image( + self, + annotation: Dict, + input_image_size: Tuple[int, int], + output_image_size: Tuple[int, int], + padding, + update_bboxes, + ) -> Dict: + """ + Update the annotation for a padded image. + """ + new_annotation = {} + new_annotation["size"] = output_image_size + + for key, value in annotation.items(): + if key == "masks": + masks = value + masks = pad( + masks, + padding, + mode=PaddingMode.CONSTANT, + constant_values=0, + input_data_format=ChannelDimension.FIRST, + ) + masks = safe_squeeze(masks, 1) + new_annotation["masks"] = masks + elif key == "boxes" and update_bboxes: + boxes = value + boxes *= np.asarray( + [ + input_image_size[1] / output_image_size[1], + input_image_size[0] / output_image_size[0], + input_image_size[1] / output_image_size[1], + input_image_size[0] / output_image_size[0], + ] + ) + new_annotation["boxes"] = boxes + elif key == "size": + new_annotation["size"] = output_image_size + else: + new_annotation[key] = value + return new_annotation + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor._pad_image + def _pad_image( + self, + image: np.ndarray, + output_size: Tuple[int, int], + annotation: Optional[Dict[str, Any]] = None, + constant_values: Union[float, Iterable[float]] = 0, + data_format: Optional[ChannelDimension] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + update_bboxes: bool = True, + ) -> np.ndarray: + """ + Pad an image with zeros to the given size. + """ + input_height, input_width = get_image_size(image, channel_dim=input_data_format) + output_height, output_width = output_size + + pad_bottom = output_height - input_height + pad_right = output_width - input_width + padding = ((0, pad_bottom), (0, pad_right)) + padded_image = pad( + image, + padding, + mode=PaddingMode.CONSTANT, + constant_values=constant_values, + data_format=data_format, + input_data_format=input_data_format, + ) + if annotation is not None: + annotation = self._update_annotation_for_padded_image( + annotation, (input_height, input_width), (output_height, output_width), padding, update_bboxes + ) + return padded_image, annotation + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.pad + def pad( + self, + images: List[np.ndarray], + annotations: Optional[Union[AnnotationType, List[AnnotationType]]] = None, + constant_values: Union[float, Iterable[float]] = 0, + return_pixel_mask: bool = True, + return_tensors: Optional[Union[str, TensorType]] = None, + data_format: Optional[ChannelDimension] = None, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + update_bboxes: bool = True, + ) -> BatchFeature: + """ + Pads a batch of images to the bottom and right of the image with zeros to the size of largest height and width + in the batch and optionally returns their corresponding pixel mask. + + Args: + images (List[`np.ndarray`]): + Images to pad. + annotations (`AnnotationType` or `List[AnnotationType]`, *optional*): + Annotations to transform according to the padding that is applied to the images. + constant_values (`float` or `Iterable[float]`, *optional*): + The value to use for the padding if `mode` is `"constant"`. + return_pixel_mask (`bool`, *optional*, defaults to `True`): + Whether to return a pixel mask. + return_tensors (`str` or `TensorType`, *optional*): + The type of tensors to return. Can be one of: + - Unset: Return a list of `np.ndarray`. + - `TensorType.TENSORFLOW` or `'tf'`: Return a batch of type `tf.Tensor`. + - `TensorType.PYTORCH` or `'pt'`: Return a batch of type `torch.Tensor`. + - `TensorType.NUMPY` or `'np'`: Return a batch of type `np.ndarray`. + - `TensorType.JAX` or `'jax'`: Return a batch of type `jax.numpy.ndarray`. + data_format (`str` or `ChannelDimension`, *optional*): + The channel dimension format of the image. If not provided, it will be the same as the input image. + input_data_format (`ChannelDimension` or `str`, *optional*): + The channel dimension format of the input image. If not provided, it will be inferred. + update_bboxes (`bool`, *optional*, defaults to `True`): + Whether to update the bounding boxes in the annotations to match the padded images. If the + bounding boxes have not been converted to relative coordinates and `(centre_x, centre_y, width, height)` + format, the bounding boxes will not be updated. + """ + pad_size = get_max_height_width(images, input_data_format=input_data_format) + + annotation_list = annotations if annotations is not None else [None] * len(images) + padded_images = [] + padded_annotations = [] + for image, annotation in zip(images, annotation_list): + padded_image, padded_annotation = self._pad_image( + image, + pad_size, + annotation, + constant_values=constant_values, + data_format=data_format, + input_data_format=input_data_format, + update_bboxes=update_bboxes, + ) + padded_images.append(padded_image) + padded_annotations.append(padded_annotation) + + data = {"pixel_values": padded_images} + + if return_pixel_mask: + masks = [ + make_pixel_mask(image=image, output_size=pad_size, input_data_format=input_data_format) + for image in images + ] + data["pixel_mask"] = masks + + encoded_inputs = BatchFeature(data=data, tensor_type=return_tensors) + + if annotations is not None: + encoded_inputs["labels"] = [ + BatchFeature(annotation, tensor_type=return_tensors) for annotation in padded_annotations + ] + + return encoded_inputs + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.preprocess + def preprocess( + self, + images: ImageInput, + annotations: Optional[Union[AnnotationType, List[AnnotationType]]] = None, + return_segmentation_masks: bool = None, + masks_path: Optional[Union[str, pathlib.Path]] = None, + do_resize: Optional[bool] = None, + size: Optional[Dict[str, int]] = None, + resample=None, # PILImageResampling + do_rescale: Optional[bool] = None, + rescale_factor: Optional[Union[int, float]] = None, + do_normalize: Optional[bool] = None, + do_convert_annotations: Optional[bool] = None, + image_mean: Optional[Union[float, List[float]]] = None, + image_std: Optional[Union[float, List[float]]] = None, + do_pad: Optional[bool] = None, + format: Optional[Union[str, AnnotationFormat]] = None, + return_tensors: Optional[Union[TensorType, str]] = None, + data_format: Union[str, ChannelDimension] = ChannelDimension.FIRST, + input_data_format: Optional[Union[str, ChannelDimension]] = None, + **kwargs, + ) -> BatchFeature: + """ + Preprocess an image or a batch of images so that it can be used by the model. + + Args: + images (`ImageInput`): + Image or batch of images to preprocess. Expects a single or batch of images with pixel values ranging + from 0 to 255. If passing in images with pixel values between 0 and 1, set `do_rescale=False`. + annotations (`AnnotationType` or `List[AnnotationType]`, *optional*): + List of annotations associated with the image or batch of images. If annotation is for object + detection, the annotations should be a dictionary with the following keys: + - "image_id" (`int`): The image id. + - "annotations" (`List[Dict]`): List of annotations for an image. Each annotation should be a + dictionary. An image can have no annotations, in which case the list should be empty. + If annotation is for segmentation, the annotations should be a dictionary with the following keys: + - "image_id" (`int`): The image id. + - "segments_info" (`List[Dict]`): List of segments for an image. Each segment should be a dictionary. + An image can have no segments, in which case the list should be empty. + - "file_name" (`str`): The file name of the image. + return_segmentation_masks (`bool`, *optional*, defaults to self.return_segmentation_masks): + Whether to return segmentation masks. + masks_path (`str` or `pathlib.Path`, *optional*): + Path to the directory containing the segmentation masks. + do_resize (`bool`, *optional*, defaults to self.do_resize): + Whether to resize the image. + size (`Dict[str, int]`, *optional*, defaults to self.size): + Size of the image after resizing. + resample (`PILImageResampling`, *optional*, defaults to self.resample): + Resampling filter to use when resizing the image. + do_rescale (`bool`, *optional*, defaults to self.do_rescale): + Whether to rescale the image. + rescale_factor (`float`, *optional*, defaults to self.rescale_factor): + Rescale factor to use when rescaling the image. + do_normalize (`bool`, *optional*, defaults to self.do_normalize): + Whether to normalize the image. + do_convert_annotations (`bool`, *optional*, defaults to self.do_convert_annotations): + Whether to convert the annotations to the format expected by the model. Converts the bounding + boxes from the format `(top_left_x, top_left_y, width, height)` to `(center_x, center_y, width, height)` + and in relative coordinates. + image_mean (`float` or `List[float]`, *optional*, defaults to self.image_mean): + Mean to use when normalizing the image. + image_std (`float` or `List[float]`, *optional*, defaults to self.image_std): + Standard deviation to use when normalizing the image. + do_pad (`bool`, *optional*, defaults to self.do_pad): + Whether to pad the image. If `True` will pad the images in the batch to the largest image in the batch + and create a pixel mask. Padding will be applied to the bottom and right of the image with zeros. + format (`str` or `AnnotationFormat`, *optional*, defaults to self.format): + Format of the annotations. + return_tensors (`str` or `TensorType`, *optional*, defaults to self.return_tensors): + Type of tensors to return. If `None`, will return the list of images. + data_format (`ChannelDimension` or `str`, *optional*, defaults to `ChannelDimension.FIRST`): + The channel dimension format for the output image. Can be one of: + - `"channels_first"` or `ChannelDimension.FIRST`: image in (num_channels, height, width) format. + - `"channels_last"` or `ChannelDimension.LAST`: image in (height, width, num_channels) format. + - Unset: Use the channel dimension format of the input image. + input_data_format (`ChannelDimension` or `str`, *optional*): + The channel dimension format for the input image. If unset, the channel dimension format is inferred + from the input image. Can be one of: + - `"channels_first"` or `ChannelDimension.FIRST`: image in (num_channels, height, width) format. + - `"channels_last"` or `ChannelDimension.LAST`: image in (height, width, num_channels) format. + - `"none"` or `ChannelDimension.NONE`: image in (height, width) format. + """ + if "pad_and_return_pixel_mask" in kwargs: + logger.warning_once( + "The `pad_and_return_pixel_mask` argument is deprecated and will be removed in a future version, " + "use `do_pad` instead." + ) + do_pad = kwargs.pop("pad_and_return_pixel_mask") + + max_size = None + if "max_size" in kwargs: + logger.warning_once( + "The `max_size` argument is deprecated and will be removed in a future version, use" + " `size['longest_edge']` instead." + ) + size = kwargs.pop("max_size") + + do_resize = self.do_resize if do_resize is None else do_resize + size = self.size if size is None else size + size = get_size_dict(size=size, max_size=max_size, default_to_square=False) + resample = self.resample if resample is None else resample + do_rescale = self.do_rescale if do_rescale is None else do_rescale + rescale_factor = self.rescale_factor if rescale_factor is None else rescale_factor + do_normalize = self.do_normalize if do_normalize is None else do_normalize + image_mean = self.image_mean if image_mean is None else image_mean + image_std = self.image_std if image_std is None else image_std + do_convert_annotations = ( + self.do_convert_annotations if do_convert_annotations is None else do_convert_annotations + ) + do_pad = self.do_pad if do_pad is None else do_pad + format = self.format if format is None else format + + images = make_list_of_images(images) + + if not valid_images(images): + raise ValueError( + "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " + "torch.Tensor, tf.Tensor or jax.ndarray." + ) + validate_kwargs(captured_kwargs=kwargs.keys(), valid_processor_keys=self._valid_processor_keys) + + # Here, the pad() method pads to the maximum of (width, height). It does not need to be validated. + validate_preprocess_arguments( + do_rescale=do_rescale, + rescale_factor=rescale_factor, + do_normalize=do_normalize, + image_mean=image_mean, + image_std=image_std, + do_resize=do_resize, + size=size, + resample=resample, + ) + + if annotations is not None and isinstance(annotations, dict): + annotations = [annotations] + + if annotations is not None and len(images) != len(annotations): + raise ValueError( + f"The number of images ({len(images)}) and annotations ({len(annotations)}) do not match." + ) + + format = AnnotationFormat(format) + if annotations is not None: + validate_annotations(format, SUPPORTED_ANNOTATION_FORMATS, annotations) + + if ( + masks_path is not None + and format == AnnotationFormat.COCO_PANOPTIC + and not isinstance(masks_path, (pathlib.Path, str)) + ): + raise ValueError( + "The path to the directory containing the mask PNG files should be provided as a" + f" `pathlib.Path` or string object, but is {type(masks_path)} instead." + ) + + # All transformations expect numpy arrays + images = [to_numpy_array(image) for image in images] + + if is_scaled_image(images[0]) and do_rescale: + logger.warning_once( + "It looks like you are trying to rescale already rescaled images. If the input" + " images have pixel values between 0 and 1, set `do_rescale=False` to avoid rescaling them again." + ) + + if input_data_format is None: + # We assume that all images have the same channel dimension format. + input_data_format = infer_channel_dimension_format(images[0]) + + # prepare (COCO annotations as a list of Dict -> DETR target as a single Dict per image) + if annotations is not None: + prepared_images = [] + prepared_annotations = [] + for image, target in zip(images, annotations): + target = self.prepare_annotation( + image, + target, + format, + return_segmentation_masks=return_segmentation_masks, + masks_path=masks_path, + input_data_format=input_data_format, + ) + prepared_images.append(image) + prepared_annotations.append(target) + images = prepared_images + annotations = prepared_annotations + del prepared_images, prepared_annotations + + # transformations + if do_resize: + if annotations is not None: + resized_images, resized_annotations = [], [] + for image, target in zip(images, annotations): + orig_size = get_image_size(image, input_data_format) + resized_image = self.resize( + image, size=size, max_size=max_size, resample=resample, input_data_format=input_data_format + ) + resized_annotation = self.resize_annotation( + target, orig_size, get_image_size(resized_image, input_data_format) + ) + resized_images.append(resized_image) + resized_annotations.append(resized_annotation) + images = resized_images + annotations = resized_annotations + del resized_images, resized_annotations + else: + images = [ + self.resize(image, size=size, resample=resample, input_data_format=input_data_format) + for image in images + ] + + if do_rescale: + images = [self.rescale(image, rescale_factor, input_data_format=input_data_format) for image in images] + + if do_normalize: + images = [ + self.normalize(image, image_mean, image_std, input_data_format=input_data_format) for image in images + ] + + if do_convert_annotations and annotations is not None: + annotations = [ + self.normalize_annotation(annotation, get_image_size(image, input_data_format)) + for annotation, image in zip(annotations, images) + ] + + if do_pad: + # Pads images and returns their mask: {'pixel_values': ..., 'pixel_mask': ...} + encoded_inputs = self.pad( + images, + annotations=annotations, + return_pixel_mask=True, + data_format=data_format, + input_data_format=input_data_format, + update_bboxes=do_convert_annotations, + return_tensors=return_tensors, + ) + else: + images = [ + to_channel_dimension_format(image, data_format, input_channel_dim=input_data_format) + for image in images + ] + encoded_inputs = BatchFeature(data={"pixel_values": images}, tensor_type=return_tensors) + if annotations is not None: + encoded_inputs["labels"] = [ + BatchFeature(annotation, tensor_type=return_tensors) for annotation in annotations + ] + + return encoded_inputs + + # POSTPROCESSING METHODS - TODO: add support for other frameworks + def post_process(self, outputs, target_sizes): + """ + Converts the output of [`ConditionalDetrForObjectDetection`] into the format expected by the Pascal VOC format (xmin, ymin, xmax, ymax). + Only supports PyTorch. + + Args: + outputs ([`ConditionalDetrObjectDetectionOutput`]): + Raw outputs of the model. + target_sizes (`torch.Tensor` of shape `(batch_size, 2)`): + Tensor containing the size (h, w) of each image of the batch. For evaluation, this must be the original + image size (before any data augmentation). For visualization, this should be the image size after data + augment, but before padding. + Returns: + `List[Dict]`: A list of dictionaries, each dictionary containing the scores, labels and boxes for an image + in the batch as predicted by the model. + """ + logging.warning_once( + "`post_process` is deprecated and will be removed in v5 of Transformers, please use" + " `post_process_object_detection` instead, with `threshold=0.` for equivalent results.", + ) + + out_logits, out_bbox = outputs.logits, outputs.pred_boxes + + if len(out_logits) != len(target_sizes): + raise ValueError("Make sure that you pass in as many target sizes as the batch dimension of the logits") + if target_sizes.shape[1] != 2: + raise ValueError("Each element of target_sizes must contain the size (h, w) of each image of the batch") + + prob = out_logits.sigmoid() + topk_values, topk_indexes = torch.topk(prob.view(out_logits.shape[0], -1), 300, dim=1) + scores = topk_values + topk_boxes = torch.div(topk_indexes, out_logits.shape[2], rounding_mode="floor") + labels = topk_indexes % out_logits.shape[2] + boxes = center_to_corners_format(out_bbox) + boxes = torch.gather(boxes, 1, topk_boxes.unsqueeze(-1).repeat(1, 1, 4)) + + # and from relative [0, 1] to absolute [0, height] coordinates + img_h, img_w = target_sizes.unbind(1) + scale_fct = torch.stack([img_w, img_h, img_w, img_h], dim=1) + boxes = boxes * scale_fct[:, None, :] + + results = [{"scores": s, "labels": l, "boxes": b} for s, l, b in zip(scores, labels, boxes)] + + return results + + # Copied from transformers.models.deformable_detr.image_processing_deformable_detr.DeformableDetrImageProcessor.post_process_object_detection with DeformableDetr->ConditionalDetr + def post_process_object_detection( + self, outputs, threshold: float = 0.5, target_sizes: Union[TensorType, List[Tuple]] = None, top_k: int = 100 + ): + """ + Converts the raw output of [`ConditionalDetrForObjectDetection`] into final bounding boxes in (top_left_x, + top_left_y, bottom_right_x, bottom_right_y) format. Only supports PyTorch. + + Args: + outputs ([`DetrObjectDetectionOutput`]): + Raw outputs of the model. + threshold (`float`, *optional*): + Score threshold to keep object detection predictions. + target_sizes (`torch.Tensor` or `List[Tuple[int, int]]`, *optional*): + Tensor of shape `(batch_size, 2)` or list of tuples (`Tuple[int, int]`) containing the target size + (height, width) of each image in the batch. If left to None, predictions will not be resized. + top_k (`int`, *optional*, defaults to 100): + Keep only top k bounding boxes before filtering by thresholding. + + Returns: + `List[Dict]`: A list of dictionaries, each dictionary containing the scores, labels and boxes for an image + in the batch as predicted by the model. + """ + out_logits, out_bbox = outputs.logits, outputs.pred_boxes + + if target_sizes is not None: + if len(out_logits) != len(target_sizes): + raise ValueError( + "Make sure that you pass in as many target sizes as the batch dimension of the logits" + ) + + prob = out_logits.sigmoid() + prob = prob.view(out_logits.shape[0], -1) + k_value = min(top_k, prob.size(1)) + topk_values, topk_indexes = torch.topk(prob, k_value, dim=1) + scores = topk_values + topk_boxes = torch.div(topk_indexes, out_logits.shape[2], rounding_mode="floor") + labels = topk_indexes % out_logits.shape[2] + boxes = center_to_corners_format(out_bbox) + boxes = torch.gather(boxes, 1, topk_boxes.unsqueeze(-1).repeat(1, 1, 4)) + + # and from relative [0, 1] to absolute [0, height] coordinates + if target_sizes is not None: + if isinstance(target_sizes, List): + img_h = torch.Tensor([i[0] for i in target_sizes]) + img_w = torch.Tensor([i[1] for i in target_sizes]) + else: + img_h, img_w = target_sizes.unbind(1) + scale_fct = torch.stack([img_w, img_h, img_w, img_h], dim=1).to(boxes.device) + boxes = boxes * scale_fct[:, None, :] + + results = [] + for s, l, b in zip(scores, labels, boxes): + score = s[s > threshold] + label = l[s > threshold] + box = b[s > threshold] + results.append({"scores": score, "labels": label, "boxes": box}) + + return results + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.post_process_semantic_segmentation with Detr->ConditionalDetr + def post_process_semantic_segmentation(self, outputs, target_sizes: List[Tuple[int, int]] = None): + """ + Converts the output of [`ConditionalDetrForSegmentation`] into semantic segmentation maps. Only supports PyTorch. + + Args: + outputs ([`ConditionalDetrForSegmentation`]): + Raw outputs of the model. + target_sizes (`List[Tuple[int, int]]`, *optional*): + A list of tuples (`Tuple[int, int]`) containing the target size (height, width) of each image in the + batch. If unset, predictions will not be resized. + Returns: + `List[torch.Tensor]`: + A list of length `batch_size`, where each item is a semantic segmentation map of shape (height, width) + corresponding to the target_sizes entry (if `target_sizes` is specified). Each entry of each + `torch.Tensor` correspond to a semantic class id. + """ + class_queries_logits = outputs.logits # [batch_size, num_queries, num_classes+1] + masks_queries_logits = outputs.pred_masks # [batch_size, num_queries, height, width] + + # Remove the null class `[..., :-1]` + masks_classes = class_queries_logits.softmax(dim=-1)[..., :-1] + masks_probs = masks_queries_logits.sigmoid() # [batch_size, num_queries, height, width] + + # Semantic segmentation logits of shape (batch_size, num_classes, height, width) + segmentation = torch.einsum("bqc, bqhw -> bchw", masks_classes, masks_probs) + batch_size = class_queries_logits.shape[0] + + # Resize logits and compute semantic segmentation maps + if target_sizes is not None: + if batch_size != len(target_sizes): + raise ValueError( + "Make sure that you pass in as many target sizes as the batch dimension of the logits" + ) + + semantic_segmentation = [] + for idx in range(batch_size): + resized_logits = nn.functional.interpolate( + segmentation[idx].unsqueeze(dim=0), size=target_sizes[idx], mode="bilinear", align_corners=False + ) + semantic_map = resized_logits[0].argmax(dim=0) + semantic_segmentation.append(semantic_map) + else: + semantic_segmentation = segmentation.argmax(dim=1) + semantic_segmentation = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0])] + + return semantic_segmentation + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.post_process_instance_segmentation with Detr->ConditionalDetr + def post_process_instance_segmentation( + self, + outputs, + threshold: float = 0.5, + mask_threshold: float = 0.5, + overlap_mask_area_threshold: float = 0.8, + target_sizes: Optional[List[Tuple[int, int]]] = None, + return_coco_annotation: Optional[bool] = False, + ) -> List[Dict]: + """ + Converts the output of [`ConditionalDetrForSegmentation`] into instance segmentation predictions. Only supports PyTorch. + + Args: + outputs ([`ConditionalDetrForSegmentation`]): + Raw outputs of the model. + threshold (`float`, *optional*, defaults to 0.5): + The probability score threshold to keep predicted instance masks. + mask_threshold (`float`, *optional*, defaults to 0.5): + Threshold to use when turning the predicted masks into binary values. + overlap_mask_area_threshold (`float`, *optional*, defaults to 0.8): + The overlap mask area threshold to merge or discard small disconnected parts within each binary + instance mask. + target_sizes (`List[Tuple]`, *optional*): + List of length (batch_size), where each list item (`Tuple[int, int]]`) corresponds to the requested + final size (height, width) of each prediction. If unset, predictions will not be resized. + return_coco_annotation (`bool`, *optional*): + Defaults to `False`. If set to `True`, segmentation maps are returned in COCO run-length encoding (RLE) + format. + Returns: + `List[Dict]`: A list of dictionaries, one per image, each dictionary containing two keys: + - **segmentation** -- A tensor of shape `(height, width)` where each pixel represents a `segment_id` or + `List[List]` run-length encoding (RLE) of the segmentation map if return_coco_annotation is set to + `True`. Set to `None` if no mask if found above `threshold`. + - **segments_info** -- A dictionary that contains additional information on each segment. + - **id** -- An integer representing the `segment_id`. + - **label_id** -- An integer representing the label / semantic class id corresponding to `segment_id`. + - **score** -- Prediction score of segment with `segment_id`. + """ + class_queries_logits = outputs.logits # [batch_size, num_queries, num_classes+1] + masks_queries_logits = outputs.pred_masks # [batch_size, num_queries, height, width] + + batch_size = class_queries_logits.shape[0] + num_labels = class_queries_logits.shape[-1] - 1 + + mask_probs = masks_queries_logits.sigmoid() # [batch_size, num_queries, height, width] + + # Predicted label and score of each query (batch_size, num_queries) + pred_scores, pred_labels = nn.functional.softmax(class_queries_logits, dim=-1).max(-1) + + # Loop over items in batch size + results: List[Dict[str, TensorType]] = [] + + for i in range(batch_size): + mask_probs_item, pred_scores_item, pred_labels_item = remove_low_and_no_objects( + mask_probs[i], pred_scores[i], pred_labels[i], threshold, num_labels + ) + + # No mask found + if mask_probs_item.shape[0] <= 0: + height, width = target_sizes[i] if target_sizes is not None else mask_probs_item.shape[1:] + segmentation = torch.zeros((height, width)) - 1 + results.append({"segmentation": segmentation, "segments_info": []}) + continue + + # Get segmentation map and segment information of batch item + target_size = target_sizes[i] if target_sizes is not None else None + segmentation, segments = compute_segments( + mask_probs=mask_probs_item, + pred_scores=pred_scores_item, + pred_labels=pred_labels_item, + mask_threshold=mask_threshold, + overlap_mask_area_threshold=overlap_mask_area_threshold, + label_ids_to_fuse=[], + target_size=target_size, + ) + + # Return segmentation map in run-length encoding (RLE) format + if return_coco_annotation: + segmentation = convert_segmentation_to_rle(segmentation) + + results.append({"segmentation": segmentation, "segments_info": segments}) + return results + + # Copied from transformers.models.detr.image_processing_detr.DetrImageProcessor.post_process_panoptic_segmentation with Detr->ConditionalDetr + def post_process_panoptic_segmentation( + self, + outputs, + threshold: float = 0.5, + mask_threshold: float = 0.5, + overlap_mask_area_threshold: float = 0.8, + label_ids_to_fuse: Optional[Set[int]] = None, + target_sizes: Optional[List[Tuple[int, int]]] = None, + ) -> List[Dict]: + """ + Converts the output of [`ConditionalDetrForSegmentation`] into image panoptic segmentation predictions. Only supports + PyTorch. + + Args: + outputs ([`ConditionalDetrForSegmentation`]): + The outputs from [`ConditionalDetrForSegmentation`]. + threshold (`float`, *optional*, defaults to 0.5): + The probability score threshold to keep predicted instance masks. + mask_threshold (`float`, *optional*, defaults to 0.5): + Threshold to use when turning the predicted masks into binary values. + overlap_mask_area_threshold (`float`, *optional*, defaults to 0.8): + The overlap mask area threshold to merge or discard small disconnected parts within each binary + instance mask. + label_ids_to_fuse (`Set[int]`, *optional*): + The labels in this state will have all their instances be fused together. For instance we could say + there can only be one sky in an image, but several persons, so the label ID for sky would be in that + set, but not the one for person. + target_sizes (`List[Tuple]`, *optional*): + List of length (batch_size), where each list item (`Tuple[int, int]]`) corresponds to the requested + final size (height, width) of each prediction in batch. If unset, predictions will not be resized. + Returns: + `List[Dict]`: A list of dictionaries, one per image, each dictionary containing two keys: + - **segmentation** -- a tensor of shape `(height, width)` where each pixel represents a `segment_id` or + `None` if no mask if found above `threshold`. If `target_sizes` is specified, segmentation is resized to + the corresponding `target_sizes` entry. + - **segments_info** -- A dictionary that contains additional information on each segment. + - **id** -- an integer representing the `segment_id`. + - **label_id** -- An integer representing the label / semantic class id corresponding to `segment_id`. + - **was_fused** -- a boolean, `True` if `label_id` was in `label_ids_to_fuse`, `False` otherwise. + Multiple instances of the same class / label were fused and assigned a single `segment_id`. + - **score** -- Prediction score of segment with `segment_id`. + """ + + if label_ids_to_fuse is None: + logger.warning_once("`label_ids_to_fuse` unset. No instance will be fused.") + label_ids_to_fuse = set() + + class_queries_logits = outputs.logits # [batch_size, num_queries, num_classes+1] + masks_queries_logits = outputs.pred_masks # [batch_size, num_queries, height, width] + + batch_size = class_queries_logits.shape[0] + num_labels = class_queries_logits.shape[-1] - 1 + + mask_probs = masks_queries_logits.sigmoid() # [batch_size, num_queries, height, width] + + # Predicted label and score of each query (batch_size, num_queries) + pred_scores, pred_labels = nn.functional.softmax(class_queries_logits, dim=-1).max(-1) + + # Loop over items in batch size + results: List[Dict[str, TensorType]] = [] + + for i in range(batch_size): + mask_probs_item, pred_scores_item, pred_labels_item = remove_low_and_no_objects( + mask_probs[i], pred_scores[i], pred_labels[i], threshold, num_labels + ) + + # No mask found + if mask_probs_item.shape[0] <= 0: + height, width = target_sizes[i] if target_sizes is not None else mask_probs_item.shape[1:] + segmentation = torch.zeros((height, width)) - 1 + results.append({"segmentation": segmentation, "segments_info": []}) + continue + + # Get segmentation map and segment information of batch item + target_size = target_sizes[i] if target_sizes is not None else None + segmentation, segments = compute_segments( + mask_probs=mask_probs_item, + pred_scores=pred_scores_item, + pred_labels=pred_labels_item, + mask_threshold=mask_threshold, + overlap_mask_area_threshold=overlap_mask_area_threshold, + label_ids_to_fuse=label_ids_to_fuse, + target_size=target_size, + ) + + results.append({"segmentation": segmentation, "segments_info": segments}) + return results diff --git a/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/__init__.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/__init__.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..5bb665890cc76a00c9f27e152ca17fda8ff2843d Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/__init__.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/configuration_llama.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/configuration_llama.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..ce6df312fa3ea85d589b0197a5099ee6fae9b041 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/configuration_llama.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/convert_llama_weights_to_hf.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/convert_llama_weights_to_hf.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..99270ebd606130c72452c933ca991e121c9ee29b Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/convert_llama_weights_to_hf.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/modeling_flax_llama.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/modeling_flax_llama.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..583478d6b1f5e55877b28ee00ccd29a0eb8c5eca Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/modeling_flax_llama.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/modeling_llama.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/modeling_llama.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..a1cea9797e9f532ba08aedd9477b603ccd5a0094 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/modeling_llama.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/tokenization_llama.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/tokenization_llama.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..9c1dd8cbfc21619c87ac137102aa96f745d5fc54 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/tokenization_llama.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/tokenization_llama_fast.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/tokenization_llama_fast.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..16aa6d7fe145991547a44f66a47d8244625a2b98 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/llama/__pycache__/tokenization_llama_fast.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/__init__.py b/venv/lib/python3.10/site-packages/transformers/models/rembert/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..98e8e2254dcfa96e80fc8c8a504a767ae6a36b09 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/rembert/__init__.py @@ -0,0 +1,150 @@ +# Copyright 2020 The HuggingFace Team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +from typing import TYPE_CHECKING + +from ...utils import ( + OptionalDependencyNotAvailable, + _LazyModule, + is_sentencepiece_available, + is_tf_available, + is_tokenizers_available, + is_torch_available, +) + + +_import_structure = { + "configuration_rembert": ["REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "RemBertConfig", "RemBertOnnxConfig"] +} + +try: + if not is_sentencepiece_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["tokenization_rembert"] = ["RemBertTokenizer"] + +try: + if not is_tokenizers_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["tokenization_rembert_fast"] = ["RemBertTokenizerFast"] + +try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["modeling_rembert"] = [ + "REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST", + "RemBertForCausalLM", + "RemBertForMaskedLM", + "RemBertForMultipleChoice", + "RemBertForQuestionAnswering", + "RemBertForSequenceClassification", + "RemBertForTokenClassification", + "RemBertLayer", + "RemBertModel", + "RemBertPreTrainedModel", + "load_tf_weights_in_rembert", + ] + + +try: + if not is_tf_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["modeling_tf_rembert"] = [ + "TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST", + "TFRemBertForCausalLM", + "TFRemBertForMaskedLM", + "TFRemBertForMultipleChoice", + "TFRemBertForQuestionAnswering", + "TFRemBertForSequenceClassification", + "TFRemBertForTokenClassification", + "TFRemBertLayer", + "TFRemBertModel", + "TFRemBertPreTrainedModel", + ] + + +if TYPE_CHECKING: + from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig + + try: + if not is_sentencepiece_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .tokenization_rembert import RemBertTokenizer + + try: + if not is_tokenizers_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .tokenization_rembert_fast import RemBertTokenizerFast + + try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .modeling_rembert import ( + REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, + RemBertForCausalLM, + RemBertForMaskedLM, + RemBertForMultipleChoice, + RemBertForQuestionAnswering, + RemBertForSequenceClassification, + RemBertForTokenClassification, + RemBertLayer, + RemBertModel, + RemBertPreTrainedModel, + load_tf_weights_in_rembert, + ) + + try: + if not is_tf_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .modeling_tf_rembert import ( + TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, + TFRemBertForCausalLM, + TFRemBertForMaskedLM, + TFRemBertForMultipleChoice, + TFRemBertForQuestionAnswering, + TFRemBertForSequenceClassification, + TFRemBertForTokenClassification, + TFRemBertLayer, + TFRemBertModel, + TFRemBertPreTrainedModel, + ) + + +else: + import sys + + sys.modules[__name__] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__) diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/__init__.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/__init__.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..7a9721ea827a7b450555acf4f9076603f666a821 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/__init__.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/configuration_rembert.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/configuration_rembert.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..7b947f0213455932d292409e26f39c1ee18e2234 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/configuration_rembert.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/convert_rembert_tf_checkpoint_to_pytorch.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/convert_rembert_tf_checkpoint_to_pytorch.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..4f0996284b78aa69e056215f6b87fa87fa712732 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/convert_rembert_tf_checkpoint_to_pytorch.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/modeling_rembert.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/modeling_rembert.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..0ee28f5e13250646b83dcaddc2a4d5a0295a52e6 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/modeling_rembert.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/modeling_tf_rembert.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/modeling_tf_rembert.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..9d7682c5c90b6f612ef6b79aa0970e7f9fe836d2 Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/modeling_tf_rembert.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/tokenization_rembert.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/tokenization_rembert.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..028f1d4cc32f3b0e60003760d63c77892938ce3e Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/tokenization_rembert.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/tokenization_rembert_fast.cpython-310.pyc b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/tokenization_rembert_fast.cpython-310.pyc new file mode 100644 index 0000000000000000000000000000000000000000..1b4730c10709bb2b2f36bf04ff617c398c2ad66e Binary files /dev/null and b/venv/lib/python3.10/site-packages/transformers/models/rembert/__pycache__/tokenization_rembert_fast.cpython-310.pyc differ diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/configuration_rembert.py b/venv/lib/python3.10/site-packages/transformers/models/rembert/configuration_rembert.py new file mode 100644 index 0000000000000000000000000000000000000000..fa51a79f6012b6ff733ecfffdc93f43a83ccacec --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/rembert/configuration_rembert.py @@ -0,0 +1,161 @@ +# coding=utf-8 +# Copyright The HuggingFace Team and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" RemBERT model configuration""" +from collections import OrderedDict +from typing import Mapping + +from ...configuration_utils import PretrainedConfig +from ...onnx import OnnxConfig +from ...utils import logging + + +logger = logging.get_logger(__name__) + + +from ..deprecated._archive_maps import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP # noqa: F401, E402 + + +class RemBertConfig(PretrainedConfig): + r""" + This is the configuration class to store the configuration of a [`RemBertModel`]. It is used to instantiate an + RemBERT model according to the specified arguments, defining the model architecture. Instantiating a configuration + with the defaults will yield a similar configuration to that of the RemBERT + [google/rembert](https://huggingface.co/google/rembert) architecture. + + Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the + documentation from [`PretrainedConfig`] for more information. + + + Args: + vocab_size (`int`, *optional*, defaults to 250300): + Vocabulary size of the RemBERT model. Defines the number of different tokens that can be represented by the + `inputs_ids` passed when calling [`RemBertModel`] or [`TFRemBertModel`]. Vocabulary size of the model. + Defines the different tokens that can be represented by the *inputs_ids* passed to the forward method of + [`RemBertModel`]. + hidden_size (`int`, *optional*, defaults to 1152): + Dimensionality of the encoder layers and the pooler layer. + num_hidden_layers (`int`, *optional*, defaults to 32): + Number of hidden layers in the Transformer encoder. + num_attention_heads (`int`, *optional*, defaults to 18): + Number of attention heads for each attention layer in the Transformer encoder. + input_embedding_size (`int`, *optional*, defaults to 256): + Dimensionality of the input embeddings. + output_embedding_size (`int`, *optional*, defaults to 1664): + Dimensionality of the output embeddings. + intermediate_size (`int`, *optional*, defaults to 4608): + Dimensionality of the "intermediate" (i.e., feed-forward) layer in the Transformer encoder. + hidden_act (`str` or `function`, *optional*, defaults to `"gelu"`): + The non-linear activation function (function or string) in the encoder and pooler. If string, `"gelu"`, + `"relu"`, `"selu"` and `"gelu_new"` are supported. + hidden_dropout_prob (`float`, *optional*, defaults to 0): + The dropout probability for all fully connected layers in the embeddings, encoder, and pooler. + attention_probs_dropout_prob (`float`, *optional*, defaults to 0): + The dropout ratio for the attention probabilities. + classifier_dropout_prob (`float`, *optional*, defaults to 0.1): + The dropout ratio for the classifier layer when fine-tuning. + max_position_embeddings (`int`, *optional*, defaults to 512): + The maximum sequence length that this model might ever be used with. Typically set this to something large + just in case (e.g., 512 or 1024 or 2048). + type_vocab_size (`int`, *optional*, defaults to 2): + The vocabulary size of the `token_type_ids` passed when calling [`RemBertModel`] or [`TFRemBertModel`]. + initializer_range (`float`, *optional*, defaults to 0.02): + The standard deviation of the truncated_normal_initializer for initializing all weight matrices. + layer_norm_eps (`float`, *optional*, defaults to 1e-12): + The epsilon used by the layer normalization layers. + is_decoder (`bool`, *optional*, defaults to `False`): + Whether the model is used as a decoder or not. If `False`, the model is used as an encoder. + use_cache (`bool`, *optional*, defaults to `True`): + Whether or not the model should return the last key/values attentions (not used by all models). Only + relevant if `config.is_decoder=True`. + + Example: + + ```python + >>> from transformers import RemBertModel, RemBertConfig + + >>> # Initializing a RemBERT rembert style configuration + >>> configuration = RemBertConfig() + + >>> # Initializing a model from the rembert style configuration + >>> model = RemBertModel(configuration) + + >>> # Accessing the model configuration + >>> configuration = model.config + ```""" + + model_type = "rembert" + + def __init__( + self, + vocab_size=250300, + hidden_size=1152, + num_hidden_layers=32, + num_attention_heads=18, + input_embedding_size=256, + output_embedding_size=1664, + intermediate_size=4608, + hidden_act="gelu", + hidden_dropout_prob=0.0, + attention_probs_dropout_prob=0.0, + classifier_dropout_prob=0.1, + max_position_embeddings=512, + type_vocab_size=2, + initializer_range=0.02, + layer_norm_eps=1e-12, + use_cache=True, + pad_token_id=0, + bos_token_id=312, + eos_token_id=313, + **kwargs, + ): + super().__init__(pad_token_id=pad_token_id, bos_token_id=bos_token_id, eos_token_id=eos_token_id, **kwargs) + + self.vocab_size = vocab_size + self.input_embedding_size = input_embedding_size + self.output_embedding_size = output_embedding_size + self.max_position_embeddings = max_position_embeddings + self.hidden_size = hidden_size + self.num_hidden_layers = num_hidden_layers + self.num_attention_heads = num_attention_heads + self.intermediate_size = intermediate_size + self.hidden_act = hidden_act + self.hidden_dropout_prob = hidden_dropout_prob + self.attention_probs_dropout_prob = attention_probs_dropout_prob + self.classifier_dropout_prob = classifier_dropout_prob + self.initializer_range = initializer_range + self.type_vocab_size = type_vocab_size + self.layer_norm_eps = layer_norm_eps + self.use_cache = use_cache + self.tie_word_embeddings = False + + +class RemBertOnnxConfig(OnnxConfig): + @property + def inputs(self) -> Mapping[str, Mapping[int, str]]: + if self.task == "multiple-choice": + dynamic_axis = {0: "batch", 1: "choice", 2: "sequence"} + else: + dynamic_axis = {0: "batch", 1: "sequence"} + return OrderedDict( + [ + ("input_ids", dynamic_axis), + ("attention_mask", dynamic_axis), + ("token_type_ids", dynamic_axis), + ] + ) + + @property + def atol_for_validation(self) -> float: + return 1e-4 diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/convert_rembert_tf_checkpoint_to_pytorch.py b/venv/lib/python3.10/site-packages/transformers/models/rembert/convert_rembert_tf_checkpoint_to_pytorch.py new file mode 100644 index 0000000000000000000000000000000000000000..4c3d53e789de011fa1933bac4904075c44965a08 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/rembert/convert_rembert_tf_checkpoint_to_pytorch.py @@ -0,0 +1,63 @@ +# coding=utf-8 +# Copyright 2018 The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Convert RemBERT checkpoint.""" + + +import argparse + +import torch + +from transformers import RemBertConfig, RemBertModel, load_tf_weights_in_rembert +from transformers.utils import logging + + +logging.set_verbosity_info() + + +def convert_rembert_tf_checkpoint_to_pytorch(tf_checkpoint_path, bert_config_file, pytorch_dump_path): + # Initialise PyTorch model + config = RemBertConfig.from_json_file(bert_config_file) + print("Building PyTorch model from configuration: {}".format(str(config))) + model = RemBertModel(config) + + # Load weights from tf checkpoint + load_tf_weights_in_rembert(model, config, tf_checkpoint_path) + + # Save pytorch-model + print("Save PyTorch model to {}".format(pytorch_dump_path)) + torch.save(model.state_dict(), pytorch_dump_path) + + +if __name__ == "__main__": + parser = argparse.ArgumentParser() + # Required parameters + parser.add_argument( + "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." + ) + parser.add_argument( + "--rembert_config_file", + default=None, + type=str, + required=True, + help=( + "The config json file corresponding to the pre-trained RemBERT model. \n" + "This specifies the model architecture." + ), + ) + parser.add_argument( + "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." + ) + args = parser.parse_args() + convert_rembert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.rembert_config_file, args.pytorch_dump_path) diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/modeling_rembert.py b/venv/lib/python3.10/site-packages/transformers/models/rembert/modeling_rembert.py new file mode 100644 index 0000000000000000000000000000000000000000..9c04ed10b8e9d885568927894d8c84ae81e9e267 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/rembert/modeling_rembert.py @@ -0,0 +1,1526 @@ +# coding=utf-8 +# Copyright 2021 The HuggingFace Team The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" PyTorch RemBERT model.""" + + +import math +import os +from typing import Optional, Tuple, Union + +import torch +import torch.utils.checkpoint +from torch import nn +from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss + +from ...activations import ACT2FN +from ...modeling_outputs import ( + BaseModelOutputWithPastAndCrossAttentions, + BaseModelOutputWithPoolingAndCrossAttentions, + CausalLMOutputWithCrossAttentions, + MaskedLMOutput, + MultipleChoiceModelOutput, + QuestionAnsweringModelOutput, + SequenceClassifierOutput, + TokenClassifierOutput, +) +from ...modeling_utils import PreTrainedModel +from ...pytorch_utils import apply_chunking_to_forward, find_pruneable_heads_and_indices, prune_linear_layer +from ...utils import ( + add_code_sample_docstrings, + add_start_docstrings, + add_start_docstrings_to_model_forward, + logging, + replace_return_docstrings, +) +from .configuration_rembert import RemBertConfig + + +logger = logging.get_logger(__name__) + +_CONFIG_FOR_DOC = "RemBertConfig" +_CHECKPOINT_FOR_DOC = "google/rembert" + + +from ..deprecated._archive_maps import REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST # noqa: F401, E402 + + +def load_tf_weights_in_rembert(model, config, tf_checkpoint_path): + """Load tf checkpoints in a pytorch model.""" + try: + import re + + import numpy as np + import tensorflow as tf + except ImportError: + logger.error( + "Loading a TensorFlow model in PyTorch, requires TensorFlow to be installed. Please see " + "https://www.tensorflow.org/install/ for installation instructions." + ) + raise + tf_path = os.path.abspath(tf_checkpoint_path) + logger.info(f"Converting TensorFlow checkpoint from {tf_path}") + # Load weights from TF model + init_vars = tf.train.list_variables(tf_path) + names = [] + arrays = [] + for name, shape in init_vars: + # Checkpoint is 12Gb, save memory by not loading useless variables + # Output embedding and cls are reset at classification time + if any(deny in name for deny in ("adam_v", "adam_m", "output_embedding", "cls")): + # logger.info("Skipping loading of %s", name) + continue + logger.info(f"Loading TF weight {name} with shape {shape}") + array = tf.train.load_variable(tf_path, name) + names.append(name) + arrays.append(array) + + for name, array in zip(names, arrays): + # Replace prefix with right one + name = name.replace("bert/", "rembert/") + # The pooler is a linear layer + # name = name.replace("pooler/dense", "pooler") + + name = name.split("/") + # adam_v and adam_m are variables used in AdamWeightDecayOptimizer to calculated m and v + # which are not required for using pretrained model + if any( + n in ["adam_v", "adam_m", "AdamWeightDecayOptimizer", "AdamWeightDecayOptimizer_1", "global_step"] + for n in name + ): + logger.info(f"Skipping {'/'.join(name)}") + continue + pointer = model + for m_name in name: + if re.fullmatch(r"[A-Za-z]+_\d+", m_name): + scope_names = re.split(r"_(\d+)", m_name) + else: + scope_names = [m_name] + if scope_names[0] == "kernel" or scope_names[0] == "gamma": + pointer = getattr(pointer, "weight") + elif scope_names[0] == "output_bias" or scope_names[0] == "beta": + pointer = getattr(pointer, "bias") + elif scope_names[0] == "output_weights": + pointer = getattr(pointer, "weight") + elif scope_names[0] == "squad": + pointer = getattr(pointer, "classifier") + else: + try: + pointer = getattr(pointer, scope_names[0]) + except AttributeError: + logger.info("Skipping {}".format("/".join(name))) + continue + if len(scope_names) >= 2: + num = int(scope_names[1]) + pointer = pointer[num] + if m_name[-11:] == "_embeddings": + pointer = getattr(pointer, "weight") + elif m_name == "kernel": + array = np.transpose(array) + try: + if pointer.shape != array.shape: + raise ValueError(f"Pointer shape {pointer.shape} and array shape {array.shape} mismatched") + except AssertionError as e: + e.args += (pointer.shape, array.shape) + raise + logger.info(f"Initialize PyTorch weight {name}") + pointer.data = torch.from_numpy(array) + return model + + +class RemBertEmbeddings(nn.Module): + """Construct the embeddings from word, position and token_type embeddings.""" + + def __init__(self, config): + super().__init__() + self.word_embeddings = nn.Embedding( + config.vocab_size, config.input_embedding_size, padding_idx=config.pad_token_id + ) + self.position_embeddings = nn.Embedding(config.max_position_embeddings, config.input_embedding_size) + self.token_type_embeddings = nn.Embedding(config.type_vocab_size, config.input_embedding_size) + + # self.LayerNorm is not snake-cased to stick with TensorFlow model variable name and be able to load + # any TensorFlow checkpoint file + self.LayerNorm = nn.LayerNorm(config.input_embedding_size, eps=config.layer_norm_eps) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + # position_ids (1, len position emb) is contiguous in memory and exported when serialized + self.register_buffer( + "position_ids", torch.arange(config.max_position_embeddings).expand((1, -1)), persistent=False + ) + + def forward( + self, + input_ids: Optional[torch.LongTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.LongTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + past_key_values_length: int = 0, + ) -> torch.Tensor: + if input_ids is not None: + input_shape = input_ids.size() + else: + input_shape = inputs_embeds.size()[:-1] + + seq_length = input_shape[1] + + if position_ids is None: + position_ids = self.position_ids[:, past_key_values_length : seq_length + past_key_values_length] + + if token_type_ids is None: + token_type_ids = torch.zeros(input_shape, dtype=torch.long, device=self.position_ids.device) + + if inputs_embeds is None: + inputs_embeds = self.word_embeddings(input_ids) + token_type_embeddings = self.token_type_embeddings(token_type_ids) + + embeddings = inputs_embeds + token_type_embeddings + position_embeddings = self.position_embeddings(position_ids) + embeddings += position_embeddings + embeddings = self.LayerNorm(embeddings) + embeddings = self.dropout(embeddings) + return embeddings + + +# Copied from transformers.models.bert.modeling_bert.BertPooler with Bert->RemBert +class RemBertPooler(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.hidden_size) + self.activation = nn.Tanh() + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + # We "pool" the model by simply taking the hidden state corresponding + # to the first token. + first_token_tensor = hidden_states[:, 0] + pooled_output = self.dense(first_token_tensor) + pooled_output = self.activation(pooled_output) + return pooled_output + + +class RemBertSelfAttention(nn.Module): + def __init__(self, config): + super().__init__() + if config.hidden_size % config.num_attention_heads != 0 and not hasattr(config, "embedding_size"): + raise ValueError( + f"The hidden size ({config.hidden_size}) is not a multiple of the number of attention " + f"heads ({config.num_attention_heads})" + ) + + self.num_attention_heads = config.num_attention_heads + self.attention_head_size = int(config.hidden_size / config.num_attention_heads) + self.all_head_size = self.num_attention_heads * self.attention_head_size + + self.query = nn.Linear(config.hidden_size, self.all_head_size) + self.key = nn.Linear(config.hidden_size, self.all_head_size) + self.value = nn.Linear(config.hidden_size, self.all_head_size) + + self.dropout = nn.Dropout(config.attention_probs_dropout_prob) + + self.is_decoder = config.is_decoder + + def transpose_for_scores(self, x): + new_x_shape = x.size()[:-1] + (self.num_attention_heads, self.attention_head_size) + x = x.view(*new_x_shape) + return x.permute(0, 2, 1, 3) + + def forward( + self, + hidden_states: torch.Tensor, + attention_mask: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_value: Tuple[Tuple[torch.FloatTensor]] = None, + output_attentions: bool = False, + ) -> Tuple: + mixed_query_layer = self.query(hidden_states) + + # If this is instantiated as a cross-attention module, the keys + # and values come from an encoder; the attention mask needs to be + # such that the encoder's padding tokens are not attended to. + is_cross_attention = encoder_hidden_states is not None + + if is_cross_attention and past_key_value is not None: + # reuse k,v, cross_attentions + key_layer = past_key_value[0] + value_layer = past_key_value[1] + attention_mask = encoder_attention_mask + elif is_cross_attention: + key_layer = self.transpose_for_scores(self.key(encoder_hidden_states)) + value_layer = self.transpose_for_scores(self.value(encoder_hidden_states)) + attention_mask = encoder_attention_mask + elif past_key_value is not None: + key_layer = self.transpose_for_scores(self.key(hidden_states)) + value_layer = self.transpose_for_scores(self.value(hidden_states)) + key_layer = torch.cat([past_key_value[0], key_layer], dim=2) + value_layer = torch.cat([past_key_value[1], value_layer], dim=2) + else: + key_layer = self.transpose_for_scores(self.key(hidden_states)) + value_layer = self.transpose_for_scores(self.value(hidden_states)) + + query_layer = self.transpose_for_scores(mixed_query_layer) + + if self.is_decoder: + # if cross_attention save Tuple(torch.Tensor, torch.Tensor) of all cross attention key/value_states. + # Further calls to cross_attention layer can then reuse all cross-attention + # key/value_states (first "if" case) + # if uni-directional self-attention (decoder) save Tuple(torch.Tensor, torch.Tensor) of + # all previous decoder key/value_states. Further calls to uni-directional self-attention + # can concat previous decoder key/value_states to current projected key/value_states (third "elif" case) + # if encoder bi-directional self-attention `past_key_value` is always `None` + past_key_value = (key_layer, value_layer) + + # Take the dot product between "query" and "key" to get the raw attention scores. + attention_scores = torch.matmul(query_layer, key_layer.transpose(-1, -2)) + + attention_scores = attention_scores / math.sqrt(self.attention_head_size) + if attention_mask is not None: + # Apply the attention mask is (precomputed for all layers in RemBertModel forward() function) + attention_scores = attention_scores + attention_mask + + # Normalize the attention scores to probabilities. + attention_probs = nn.functional.softmax(attention_scores, dim=-1) + + # This is actually dropping out entire tokens to attend to, which might + # seem a bit unusual, but is taken from the original Transformer paper. + attention_probs = self.dropout(attention_probs) + + # Mask heads if we want to + if head_mask is not None: + attention_probs = attention_probs * head_mask + + context_layer = torch.matmul(attention_probs, value_layer) + + context_layer = context_layer.permute(0, 2, 1, 3).contiguous() + new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) + context_layer = context_layer.view(*new_context_layer_shape) + + outputs = (context_layer, attention_probs) if output_attentions else (context_layer,) + + if self.is_decoder: + outputs = outputs + (past_key_value,) + return outputs + + +# Copied from transformers.models.bert.modeling_bert.BertSelfOutput with Bert->RemBert +class RemBertSelfOutput(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.hidden_size) + self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.dropout(hidden_states) + hidden_states = self.LayerNorm(hidden_states + input_tensor) + return hidden_states + + +class RemBertAttention(nn.Module): + def __init__(self, config): + super().__init__() + self.self = RemBertSelfAttention(config) + self.output = RemBertSelfOutput(config) + self.pruned_heads = set() + + # Copied from transformers.models.bert.modeling_bert.BertAttention.prune_heads + def prune_heads(self, heads): + if len(heads) == 0: + return + heads, index = find_pruneable_heads_and_indices( + heads, self.self.num_attention_heads, self.self.attention_head_size, self.pruned_heads + ) + + # Prune linear layers + self.self.query = prune_linear_layer(self.self.query, index) + self.self.key = prune_linear_layer(self.self.key, index) + self.self.value = prune_linear_layer(self.self.value, index) + self.output.dense = prune_linear_layer(self.output.dense, index, dim=1) + + # Update hyper params and store pruned heads + self.self.num_attention_heads = self.self.num_attention_heads - len(heads) + self.self.all_head_size = self.self.attention_head_size * self.self.num_attention_heads + self.pruned_heads = self.pruned_heads.union(heads) + + # Copied from transformers.models.bert.modeling_bert.BertAttention.forward + def forward( + self, + hidden_states: torch.Tensor, + attention_mask: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_value: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + output_attentions: Optional[bool] = False, + ) -> Tuple[torch.Tensor]: + self_outputs = self.self( + hidden_states, + attention_mask, + head_mask, + encoder_hidden_states, + encoder_attention_mask, + past_key_value, + output_attentions, + ) + attention_output = self.output(self_outputs[0], hidden_states) + outputs = (attention_output,) + self_outputs[1:] # add attentions if we output them + return outputs + + +# Copied from transformers.models.bert.modeling_bert.BertIntermediate with Bert->RemBert +class RemBertIntermediate(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.intermediate_size) + if isinstance(config.hidden_act, str): + self.intermediate_act_fn = ACT2FN[config.hidden_act] + else: + self.intermediate_act_fn = config.hidden_act + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.intermediate_act_fn(hidden_states) + return hidden_states + + +# Copied from transformers.models.bert.modeling_bert.BertOutput with Bert->RemBert +class RemBertOutput(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.intermediate_size, config.hidden_size) + self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.dropout(hidden_states) + hidden_states = self.LayerNorm(hidden_states + input_tensor) + return hidden_states + + +class RemBertLayer(nn.Module): + def __init__(self, config): + super().__init__() + self.chunk_size_feed_forward = config.chunk_size_feed_forward + self.seq_len_dim = 1 + self.attention = RemBertAttention(config) + self.is_decoder = config.is_decoder + self.add_cross_attention = config.add_cross_attention + if self.add_cross_attention: + if not self.is_decoder: + raise ValueError(f"{self} should be used as a decoder model if cross attention is added") + self.crossattention = RemBertAttention(config) + self.intermediate = RemBertIntermediate(config) + self.output = RemBertOutput(config) + + # Copied from transformers.models.bert.modeling_bert.BertLayer.forward + def forward( + self, + hidden_states: torch.Tensor, + attention_mask: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_value: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + output_attentions: Optional[bool] = False, + ) -> Tuple[torch.Tensor]: + # decoder uni-directional self-attention cached key/values tuple is at positions 1,2 + self_attn_past_key_value = past_key_value[:2] if past_key_value is not None else None + self_attention_outputs = self.attention( + hidden_states, + attention_mask, + head_mask, + output_attentions=output_attentions, + past_key_value=self_attn_past_key_value, + ) + attention_output = self_attention_outputs[0] + + # if decoder, the last output is tuple of self-attn cache + if self.is_decoder: + outputs = self_attention_outputs[1:-1] + present_key_value = self_attention_outputs[-1] + else: + outputs = self_attention_outputs[1:] # add self attentions if we output attention weights + + cross_attn_present_key_value = None + if self.is_decoder and encoder_hidden_states is not None: + if not hasattr(self, "crossattention"): + raise ValueError( + f"If `encoder_hidden_states` are passed, {self} has to be instantiated with cross-attention layers" + " by setting `config.add_cross_attention=True`" + ) + + # cross_attn cached key/values tuple is at positions 3,4 of past_key_value tuple + cross_attn_past_key_value = past_key_value[-2:] if past_key_value is not None else None + cross_attention_outputs = self.crossattention( + attention_output, + attention_mask, + head_mask, + encoder_hidden_states, + encoder_attention_mask, + cross_attn_past_key_value, + output_attentions, + ) + attention_output = cross_attention_outputs[0] + outputs = outputs + cross_attention_outputs[1:-1] # add cross attentions if we output attention weights + + # add cross-attn cache to positions 3,4 of present_key_value tuple + cross_attn_present_key_value = cross_attention_outputs[-1] + present_key_value = present_key_value + cross_attn_present_key_value + + layer_output = apply_chunking_to_forward( + self.feed_forward_chunk, self.chunk_size_feed_forward, self.seq_len_dim, attention_output + ) + outputs = (layer_output,) + outputs + + # if decoder, return the attn key/values as the last output + if self.is_decoder: + outputs = outputs + (present_key_value,) + + return outputs + + # Copied from transformers.models.bert.modeling_bert.BertLayer.feed_forward_chunk + def feed_forward_chunk(self, attention_output): + intermediate_output = self.intermediate(attention_output) + layer_output = self.output(intermediate_output, attention_output) + return layer_output + + +class RemBertEncoder(nn.Module): + def __init__(self, config): + super().__init__() + self.config = config + + self.embedding_hidden_mapping_in = nn.Linear(config.input_embedding_size, config.hidden_size) + self.layer = nn.ModuleList([RemBertLayer(config) for _ in range(config.num_hidden_layers)]) + self.gradient_checkpointing = False + + def forward( + self, + hidden_states: torch.Tensor, + attention_mask: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + use_cache: Optional[bool] = None, + output_attentions: bool = False, + output_hidden_states: bool = False, + return_dict: bool = True, + ) -> Union[Tuple, BaseModelOutputWithPastAndCrossAttentions]: + if self.gradient_checkpointing and self.training: + if use_cache: + logger.warning_once( + "`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`..." + ) + use_cache = False + hidden_states = self.embedding_hidden_mapping_in(hidden_states) + all_hidden_states = () if output_hidden_states else None + all_self_attentions = () if output_attentions else None + all_cross_attentions = () if output_attentions and self.config.add_cross_attention else None + + next_decoder_cache = () if use_cache else None + for i, layer_module in enumerate(self.layer): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + layer_head_mask = head_mask[i] if head_mask is not None else None + past_key_value = past_key_values[i] if past_key_values is not None else None + + if self.gradient_checkpointing and self.training: + layer_outputs = self._gradient_checkpointing_func( + layer_module.__call__, + hidden_states, + attention_mask, + layer_head_mask, + encoder_hidden_states, + encoder_attention_mask, + past_key_value, + output_attentions, + ) + else: + layer_outputs = layer_module( + hidden_states, + attention_mask, + layer_head_mask, + encoder_hidden_states, + encoder_attention_mask, + past_key_value, + output_attentions, + ) + + hidden_states = layer_outputs[0] + if use_cache: + next_decoder_cache += (layer_outputs[-1],) + if output_attentions: + all_self_attentions = all_self_attentions + (layer_outputs[1],) + if self.config.add_cross_attention: + all_cross_attentions = all_cross_attentions + (layer_outputs[2],) + + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if not return_dict: + return tuple( + v + for v in [ + hidden_states, + next_decoder_cache, + all_hidden_states, + all_self_attentions, + all_cross_attentions, + ] + if v is not None + ) + return BaseModelOutputWithPastAndCrossAttentions( + last_hidden_state=hidden_states, + past_key_values=next_decoder_cache, + hidden_states=all_hidden_states, + attentions=all_self_attentions, + cross_attentions=all_cross_attentions, + ) + + +# Copied from transformers.models.bert.modeling_bert.BertPredictionHeadTransform with Bert->RemBert +class RemBertPredictionHeadTransform(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.hidden_size) + if isinstance(config.hidden_act, str): + self.transform_act_fn = ACT2FN[config.hidden_act] + else: + self.transform_act_fn = config.hidden_act + self.LayerNorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.transform_act_fn(hidden_states) + hidden_states = self.LayerNorm(hidden_states) + return hidden_states + + +class RemBertLMPredictionHead(nn.Module): + def __init__(self, config): + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.output_embedding_size) + self.decoder = nn.Linear(config.output_embedding_size, config.vocab_size) + self.activation = ACT2FN[config.hidden_act] + self.LayerNorm = nn.LayerNorm(config.output_embedding_size, eps=config.layer_norm_eps) + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.activation(hidden_states) + hidden_states = self.LayerNorm(hidden_states) + hidden_states = self.decoder(hidden_states) + return hidden_states + + +# Copied from transformers.models.bert.modeling_bert.BertOnlyMLMHead with Bert->RemBert +class RemBertOnlyMLMHead(nn.Module): + def __init__(self, config): + super().__init__() + self.predictions = RemBertLMPredictionHead(config) + + def forward(self, sequence_output: torch.Tensor) -> torch.Tensor: + prediction_scores = self.predictions(sequence_output) + return prediction_scores + + +class RemBertPreTrainedModel(PreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = RemBertConfig + load_tf_weights = load_tf_weights_in_rembert + base_model_prefix = "rembert" + supports_gradient_checkpointing = True + + def _init_weights(self, module): + """Initialize the weights""" + if isinstance(module, nn.Linear): + # Slightly different from the TF version which uses truncated_normal for initialization + # cf https://github.com/pytorch/pytorch/pull/5617 + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.bias is not None: + module.bias.data.zero_() + elif isinstance(module, nn.Embedding): + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.padding_idx is not None: + module.weight.data[module.padding_idx].zero_() + elif isinstance(module, nn.LayerNorm): + module.bias.data.zero_() + module.weight.data.fill_(1.0) + + +REMBERT_START_DOCSTRING = r""" + This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use + it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and + behavior. + + Parameters: + config ([`RemBertConfig`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. +""" + +REMBERT_INPUTS_DOCSTRING = r""" + Args: + input_ids (`torch.LongTensor` of shape `({0})`): + Indices of input sequence tokens in the vocabulary. + + Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and + [`PreTrainedTokenizer.__call__`] for details. + + [What are input IDs?](../glossary#input-ids) + attention_mask (`torch.FloatTensor` of shape `({0})`, *optional*): + Mask to avoid performing attention on padding token indices. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + [What are attention masks?](../glossary#attention-mask) + token_type_ids (`torch.LongTensor` of shape `({0})`, *optional*): + Segment token indices to indicate first and second portions of the inputs. Indices are selected in `[0, + 1]`: + + - 0 corresponds to a *sentence A* token, + - 1 corresponds to a *sentence B* token. + + [What are token type IDs?](../glossary#token-type-ids) + position_ids (`torch.LongTensor` of shape `({0})`, *optional*): + Indices of positions of each input sequence tokens in the position embeddings. Selected in the range `[0, + config.max_position_embeddings - 1]`. + + [What are position IDs?](../glossary#position-ids) + head_mask (`torch.FloatTensor` of shape `(num_heads,)` or `(num_layers, num_heads)`, *optional*): + Mask to nullify selected heads of the self-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + inputs_embeds (`torch.FloatTensor` of shape `({0}, hidden_size)`, *optional*): + Optionally, instead of passing `input_ids` you can choose to directly pass an embedded representation. This + is useful if you want more control over how to convert *input_ids* indices into associated vectors than the + model's internal embedding lookup matrix. + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + + +@add_start_docstrings( + "The bare RemBERT Model transformer outputting raw hidden-states without any specific head on top.", + REMBERT_START_DOCSTRING, +) +class RemBertModel(RemBertPreTrainedModel): + """ + + The model can behave as an encoder (with only self-attention) as well as a decoder, in which case a layer of + cross-attention is added between the self-attention layers, following the architecture described in [Attention is + all you need](https://arxiv.org/abs/1706.03762) by Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, + Llion Jones, Aidan N. Gomez, Lukasz Kaiser and Illia Polosukhin. + + To behave as an decoder the model needs to be initialized with the `is_decoder` argument of the configuration set + to `True`. To be used in a Seq2Seq model, the model needs to initialized with both `is_decoder` argument and + `add_cross_attention` set to `True`; an `encoder_hidden_states` is then expected as an input to the forward pass. + """ + + def __init__(self, config, add_pooling_layer=True): + super().__init__(config) + self.config = config + + self.embeddings = RemBertEmbeddings(config) + self.encoder = RemBertEncoder(config) + + self.pooler = RemBertPooler(config) if add_pooling_layer else None + + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self): + return self.embeddings.word_embeddings + + def set_input_embeddings(self, value): + self.embeddings.word_embeddings = value + + def _prune_heads(self, heads_to_prune): + """ + Prunes heads of the model. heads_to_prune: dict of {layer_num: list of heads to prune in this layer} See base + class PreTrainedModel + """ + for layer, heads in heads_to_prune.items(): + self.encoder.layer[layer].attention.prune_heads(heads) + + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=BaseModelOutputWithPastAndCrossAttentions, + config_class=_CONFIG_FOR_DOC, + ) + def forward( + self, + input_ids: torch.LongTensor = None, + attention_mask: Optional[torch.LongTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.LongTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, BaseModelOutputWithPoolingAndCrossAttentions]: + r""" + encoder_hidden_states (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the encoder. Used in the cross-attention if + the model is configured as a decoder. + encoder_attention_mask (`torch.FloatTensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on the padding token indices of the encoder input. This mask is used in + the cross-attention if the model is configured as a decoder. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + past_key_values (`tuple(tuple(torch.FloatTensor))` of length `config.n_layers` with each tuple having 4 tensors of shape `(batch_size, num_heads, sequence_length - 1, embed_size_per_head)`): + Contains precomputed key and value hidden states of the attention blocks. Can be used to speed up decoding. + If `past_key_values` are used, the user can optionally input only the last `decoder_input_ids` (those that + don't have their past key value states given to this model) of shape `(batch_size, 1)` instead of all + `decoder_input_ids` of shape `(batch_size, sequence_length)`. + use_cache (`bool`, *optional*): + If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding (see + `past_key_values`). + """ + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if self.config.is_decoder: + use_cache = use_cache if use_cache is not None else self.config.use_cache + else: + use_cache = False + + if input_ids is not None and inputs_embeds is not None: + raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time") + elif input_ids is not None: + self.warn_if_padding_and_no_attention_mask(input_ids, attention_mask) + input_shape = input_ids.size() + elif inputs_embeds is not None: + input_shape = inputs_embeds.size()[:-1] + else: + raise ValueError("You have to specify either input_ids or inputs_embeds") + + batch_size, seq_length = input_shape + device = input_ids.device if input_ids is not None else inputs_embeds.device + + # past_key_values_length + past_key_values_length = past_key_values[0][0].shape[2] if past_key_values is not None else 0 + + if attention_mask is None: + attention_mask = torch.ones(((batch_size, seq_length + past_key_values_length)), device=device) + if token_type_ids is None: + token_type_ids = torch.zeros(input_shape, dtype=torch.long, device=device) + + # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] + # ourselves in which case we just need to make it broadcastable to all heads. + extended_attention_mask: torch.Tensor = self.get_extended_attention_mask(attention_mask, input_shape) + + # If a 2D or 3D attention mask is provided for the cross-attention + # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] + if self.config.is_decoder and encoder_hidden_states is not None: + encoder_batch_size, encoder_sequence_length, _ = encoder_hidden_states.size() + encoder_hidden_shape = (encoder_batch_size, encoder_sequence_length) + if encoder_attention_mask is None: + encoder_attention_mask = torch.ones(encoder_hidden_shape, device=device) + encoder_extended_attention_mask = self.invert_attention_mask(encoder_attention_mask) + else: + encoder_extended_attention_mask = None + + # Prepare head mask if needed + # 1.0 in head_mask indicate we keep the head + # attention_probs has shape bsz x n_heads x N x N + # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] + # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] + head_mask = self.get_head_mask(head_mask, self.config.num_hidden_layers) + + embedding_output = self.embeddings( + input_ids=input_ids, + position_ids=position_ids, + token_type_ids=token_type_ids, + inputs_embeds=inputs_embeds, + past_key_values_length=past_key_values_length, + ) + encoder_outputs = self.encoder( + embedding_output, + attention_mask=extended_attention_mask, + head_mask=head_mask, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_extended_attention_mask, + past_key_values=past_key_values, + use_cache=use_cache, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + sequence_output = encoder_outputs[0] + pooled_output = self.pooler(sequence_output) if self.pooler is not None else None + + if not return_dict: + return (sequence_output, pooled_output) + encoder_outputs[1:] + + return BaseModelOutputWithPoolingAndCrossAttentions( + last_hidden_state=sequence_output, + pooler_output=pooled_output, + past_key_values=encoder_outputs.past_key_values, + hidden_states=encoder_outputs.hidden_states, + attentions=encoder_outputs.attentions, + cross_attentions=encoder_outputs.cross_attentions, + ) + + +@add_start_docstrings("""RemBERT Model with a `language modeling` head on top.""", REMBERT_START_DOCSTRING) +class RemBertForMaskedLM(RemBertPreTrainedModel): + _tied_weights_keys = ["cls.predictions.decoder.weight"] + + def __init__(self, config): + super().__init__(config) + + if config.is_decoder: + logger.warning( + "If you want to use `RemBertForMaskedLM` make sure `config.is_decoder=False` for " + "bi-directional self-attention." + ) + + self.rembert = RemBertModel(config, add_pooling_layer=False) + self.cls = RemBertOnlyMLMHead(config) + + # Initialize weights and apply final processing + self.post_init() + + def get_output_embeddings(self): + return self.cls.predictions.decoder + + def set_output_embeddings(self, new_embeddings): + self.cls.predictions.decoder = new_embeddings + + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=MaskedLMOutput, + config_class=_CONFIG_FOR_DOC, + ) + def forward( + self, + input_ids: torch.LongTensor = None, + attention_mask: Optional[torch.LongTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.LongTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + labels: Optional[torch.LongTensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, MaskedLMOutput]: + r""" + labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): + Labels for computing the masked language modeling loss. Indices should be in `[-100, 0, ..., + config.vocab_size]` (see `input_ids` docstring) Tokens with indices set to `-100` are ignored (masked), the + loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]`. + """ + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + outputs = self.rembert( + input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_attention_mask, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + sequence_output = outputs[0] + prediction_scores = self.cls(sequence_output) + + masked_lm_loss = None + if labels is not None: + loss_fct = CrossEntropyLoss() # -100 index = padding token + masked_lm_loss = loss_fct(prediction_scores.view(-1, self.config.vocab_size), labels.view(-1)) + + if not return_dict: + output = (prediction_scores,) + outputs[2:] + return ((masked_lm_loss,) + output) if masked_lm_loss is not None else output + + return MaskedLMOutput( + loss=masked_lm_loss, + logits=prediction_scores, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + def prepare_inputs_for_generation(self, input_ids, attention_mask=None, **model_kwargs): + input_shape = input_ids.shape + effective_batch_size = input_shape[0] + + # add a dummy token + assert self.config.pad_token_id is not None, "The PAD token should be defined for generation" + attention_mask = torch.cat([attention_mask, attention_mask.new_zeros((attention_mask.shape[0], 1))], dim=-1) + dummy_token = torch.full( + (effective_batch_size, 1), self.config.pad_token_id, dtype=torch.long, device=input_ids.device + ) + input_ids = torch.cat([input_ids, dummy_token], dim=1) + + return {"input_ids": input_ids, "attention_mask": attention_mask} + + +@add_start_docstrings( + """RemBERT Model with a `language modeling` head on top for CLM fine-tuning.""", REMBERT_START_DOCSTRING +) +class RemBertForCausalLM(RemBertPreTrainedModel): + _tied_weights_keys = ["cls.predictions.decoder.weight"] + + def __init__(self, config): + super().__init__(config) + + if not config.is_decoder: + logger.warning("If you want to use `RemBertForCausalLM` as a standalone, add `is_decoder=True.`") + + self.rembert = RemBertModel(config, add_pooling_layer=False) + self.cls = RemBertOnlyMLMHead(config) + + # Initialize weights and apply final processing + self.post_init() + + def get_output_embeddings(self): + return self.cls.predictions.decoder + + def set_output_embeddings(self, new_embeddings): + self.cls.predictions.decoder = new_embeddings + + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @replace_return_docstrings(output_type=CausalLMOutputWithCrossAttentions, config_class=_CONFIG_FOR_DOC) + def forward( + self, + input_ids: torch.LongTensor = None, + attention_mask: Optional[torch.LongTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.LongTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + encoder_hidden_states: Optional[torch.FloatTensor] = None, + encoder_attention_mask: Optional[torch.FloatTensor] = None, + past_key_values: Optional[Tuple[Tuple[torch.FloatTensor]]] = None, + labels: Optional[torch.LongTensor] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, CausalLMOutputWithCrossAttentions]: + r""" + encoder_hidden_states (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the encoder. Used in the cross-attention if + the model is configured as a decoder. + encoder_attention_mask (`torch.FloatTensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on the padding token indices of the encoder input. This mask is used in + the cross-attention if the model is configured as a decoder. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + past_key_values (`tuple(tuple(torch.FloatTensor))` of length `config.n_layers` with each tuple having 4 tensors of shape `(batch_size, num_heads, sequence_length - 1, embed_size_per_head)`): + Contains precomputed key and value hidden states of the attention blocks. Can be used to speed up decoding. + If `past_key_values` are used, the user can optionally input only the last `decoder_input_ids` (those that + don't have their past key value states given to this model) of shape `(batch_size, 1)` instead of all + `decoder_input_ids` of shape `(batch_size, sequence_length)`. + labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): + Labels for computing the left-to-right language modeling loss (next word prediction). Indices should be in + `[-100, 0, ..., config.vocab_size]` (see `input_ids` docstring) Tokens with indices set to `-100` are + ignored (masked), the loss is only computed for the tokens with labels n `[0, ..., config.vocab_size]`. + use_cache (`bool`, *optional*): + If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding (see + `past_key_values`). + + Returns: + + Example: + + ```python + >>> from transformers import AutoTokenizer, RemBertForCausalLM, RemBertConfig + >>> import torch + + >>> tokenizer = AutoTokenizer.from_pretrained("google/rembert") + >>> config = RemBertConfig.from_pretrained("google/rembert") + >>> config.is_decoder = True + >>> model = RemBertForCausalLM.from_pretrained("google/rembert", config=config) + + >>> inputs = tokenizer("Hello, my dog is cute", return_tensors="pt") + >>> outputs = model(**inputs) + + >>> prediction_logits = outputs.logits + ```""" + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + outputs = self.rembert( + input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_attention_mask, + past_key_values=past_key_values, + use_cache=use_cache, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + sequence_output = outputs[0] + prediction_scores = self.cls(sequence_output) + + lm_loss = None + if labels is not None: + # we are doing next-token prediction; shift prediction scores and input ids by one + shifted_prediction_scores = prediction_scores[:, :-1, :].contiguous() + labels = labels[:, 1:].contiguous() + loss_fct = CrossEntropyLoss() + lm_loss = loss_fct(shifted_prediction_scores.view(-1, self.config.vocab_size), labels.view(-1)) + + if not return_dict: + output = (prediction_scores,) + outputs[2:] + return ((lm_loss,) + output) if lm_loss is not None else output + + return CausalLMOutputWithCrossAttentions( + loss=lm_loss, + logits=prediction_scores, + past_key_values=outputs.past_key_values, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + cross_attentions=outputs.cross_attentions, + ) + + def prepare_inputs_for_generation(self, input_ids, past_key_values=None, attention_mask=None, **model_kwargs): + input_shape = input_ids.shape + + # if model is used as a decoder in encoder-decoder model, the decoder attention mask is created on the fly + if attention_mask is None: + attention_mask = input_ids.new_ones(input_shape) + + # cut decoder_input_ids if past_key_values is used + if past_key_values is not None: + past_length = past_key_values[0][0].shape[2] + + # Some generation methods already pass only the last input ID + if input_ids.shape[1] > past_length: + remove_prefix_length = past_length + else: + # Default to old behavior: keep only final ID + remove_prefix_length = input_ids.shape[1] - 1 + + input_ids = input_ids[:, remove_prefix_length:] + + return {"input_ids": input_ids, "attention_mask": attention_mask, "past_key_values": past_key_values} + + def _reorder_cache(self, past_key_values, beam_idx): + reordered_past = () + for layer_past in past_key_values: + reordered_past += ( + tuple(past_state.index_select(0, beam_idx.to(past_state.device)) for past_state in layer_past[:2]) + + layer_past[2:], + ) + return reordered_past + + +@add_start_docstrings( + """ + RemBERT Model transformer with a sequence classification/regression head on top (a linear layer on top of the + pooled output) e.g. for GLUE tasks. + """, + REMBERT_START_DOCSTRING, +) +class RemBertForSequenceClassification(RemBertPreTrainedModel): + def __init__(self, config): + super().__init__(config) + self.num_labels = config.num_labels + self.rembert = RemBertModel(config) + self.dropout = nn.Dropout(config.classifier_dropout_prob) + self.classifier = nn.Linear(config.hidden_size, config.num_labels) + + # Initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=SequenceClassifierOutput, + config_class=_CONFIG_FOR_DOC, + ) + def forward( + self, + input_ids: torch.FloatTensor = None, + attention_mask: Optional[torch.FloatTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + labels: Optional[torch.LongTensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, SequenceClassifierOutput]: + r""" + labels (`torch.LongTensor` of shape `(batch_size,)`, *optional*): + Labels for computing the sequence classification/regression loss. Indices should be in `[0, ..., + config.num_labels - 1]`. If `config.num_labels == 1` a regression loss is computed (Mean-Square loss), If + `config.num_labels > 1` a classification loss is computed (Cross-Entropy). + """ + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + outputs = self.rembert( + input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + pooled_output = outputs[1] + + pooled_output = self.dropout(pooled_output) + logits = self.classifier(pooled_output) + + loss = None + if labels is not None: + if self.config.problem_type is None: + if self.num_labels == 1: + self.config.problem_type = "regression" + elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): + self.config.problem_type = "single_label_classification" + else: + self.config.problem_type = "multi_label_classification" + + if self.config.problem_type == "regression": + loss_fct = MSELoss() + if self.num_labels == 1: + loss = loss_fct(logits.squeeze(), labels.squeeze()) + else: + loss = loss_fct(logits, labels) + elif self.config.problem_type == "single_label_classification": + loss_fct = CrossEntropyLoss() + loss = loss_fct(logits.view(-1, self.num_labels), labels.view(-1)) + elif self.config.problem_type == "multi_label_classification": + loss_fct = BCEWithLogitsLoss() + loss = loss_fct(logits, labels) + if not return_dict: + output = (logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return SequenceClassifierOutput( + loss=loss, + logits=logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + +@add_start_docstrings( + """ + RemBERT Model with a multiple choice classification head on top (a linear layer on top of the pooled output and a + softmax) e.g. for RocStories/SWAG tasks. + """, + REMBERT_START_DOCSTRING, +) +class RemBertForMultipleChoice(RemBertPreTrainedModel): + def __init__(self, config): + super().__init__(config) + + self.rembert = RemBertModel(config) + self.dropout = nn.Dropout(config.classifier_dropout_prob) + self.classifier = nn.Linear(config.hidden_size, 1) + + # Initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, num_choices, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=MultipleChoiceModelOutput, + config_class=_CONFIG_FOR_DOC, + ) + def forward( + self, + input_ids: torch.FloatTensor = None, + attention_mask: Optional[torch.FloatTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + labels: Optional[torch.LongTensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, MultipleChoiceModelOutput]: + r""" + labels (`torch.LongTensor` of shape `(batch_size,)`, *optional*): + Labels for computing the multiple choice classification loss. Indices should be in `[0, ..., + num_choices-1]` where `num_choices` is the size of the second dimension of the input tensors. (See + `input_ids` above) + """ + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + num_choices = input_ids.shape[1] if input_ids is not None else inputs_embeds.shape[1] + + input_ids = input_ids.view(-1, input_ids.size(-1)) if input_ids is not None else None + attention_mask = attention_mask.view(-1, attention_mask.size(-1)) if attention_mask is not None else None + token_type_ids = token_type_ids.view(-1, token_type_ids.size(-1)) if token_type_ids is not None else None + position_ids = position_ids.view(-1, position_ids.size(-1)) if position_ids is not None else None + inputs_embeds = ( + inputs_embeds.view(-1, inputs_embeds.size(-2), inputs_embeds.size(-1)) + if inputs_embeds is not None + else None + ) + + outputs = self.rembert( + input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + pooled_output = outputs[1] + + pooled_output = self.dropout(pooled_output) + logits = self.classifier(pooled_output) + reshaped_logits = logits.view(-1, num_choices) + + loss = None + if labels is not None: + loss_fct = CrossEntropyLoss() + loss = loss_fct(reshaped_logits, labels) + + if not return_dict: + output = (reshaped_logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return MultipleChoiceModelOutput( + loss=loss, + logits=reshaped_logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + +@add_start_docstrings( + """ + RemBERT Model with a token classification head on top (a linear layer on top of the hidden-states output) e.g. for + Named-Entity-Recognition (NER) tasks. + """, + REMBERT_START_DOCSTRING, +) +class RemBertForTokenClassification(RemBertPreTrainedModel): + def __init__(self, config): + super().__init__(config) + self.num_labels = config.num_labels + + self.rembert = RemBertModel(config, add_pooling_layer=False) + self.dropout = nn.Dropout(config.classifier_dropout_prob) + self.classifier = nn.Linear(config.hidden_size, config.num_labels) + + # Initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=TokenClassifierOutput, + config_class=_CONFIG_FOR_DOC, + ) + def forward( + self, + input_ids: torch.FloatTensor = None, + attention_mask: Optional[torch.FloatTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + labels: Optional[torch.LongTensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, TokenClassifierOutput]: + r""" + labels (`torch.LongTensor` of shape `(batch_size, sequence_length)`, *optional*): + Labels for computing the token classification loss. Indices should be in `[0, ..., config.num_labels - 1]`. + """ + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + outputs = self.rembert( + input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + sequence_output = outputs[0] + + sequence_output = self.dropout(sequence_output) + logits = self.classifier(sequence_output) + + loss = None + if labels is not None: + loss_fct = CrossEntropyLoss() + loss = loss_fct(logits.view(-1, self.num_labels), labels.view(-1)) + + if not return_dict: + output = (logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return TokenClassifierOutput( + loss=loss, + logits=logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + +@add_start_docstrings( + """ + RemBERT Model with a span classification head on top for extractive question-answering tasks like SQuAD (a linear + layers on top of the hidden-states output to compute `span start logits` and `span end logits`). + """, + REMBERT_START_DOCSTRING, +) +class RemBertForQuestionAnswering(RemBertPreTrainedModel): + def __init__(self, config): + super().__init__(config) + + self.num_labels = config.num_labels + + self.rembert = RemBertModel(config, add_pooling_layer=False) + self.qa_outputs = nn.Linear(config.hidden_size, config.num_labels) + + # Initialize weights and apply final processing + self.post_init() + + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=QuestionAnsweringModelOutput, + config_class=_CONFIG_FOR_DOC, + ) + def forward( + self, + input_ids: torch.FloatTensor = None, + attention_mask: Optional[torch.FloatTensor] = None, + token_type_ids: Optional[torch.LongTensor] = None, + position_ids: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + inputs_embeds: Optional[torch.FloatTensor] = None, + start_positions: Optional[torch.LongTensor] = None, + end_positions: Optional[torch.LongTensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, QuestionAnsweringModelOutput]: + r""" + start_positions (`torch.LongTensor` of shape `(batch_size,)`, *optional*): + Labels for position (index) of the start of the labelled span for computing the token classification loss. + Positions are clamped to the length of the sequence (`sequence_length`). Position outside of the sequence + are not taken into account for computing the loss. + end_positions (`torch.LongTensor` of shape `(batch_size,)`, *optional*): + Labels for position (index) of the end of the labelled span for computing the token classification loss. + Positions are clamped to the length of the sequence (`sequence_length`). Position outside of the sequence + are not taken into account for computing the loss. + """ + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + outputs = self.rembert( + input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + sequence_output = outputs[0] + + logits = self.qa_outputs(sequence_output) + start_logits, end_logits = logits.split(1, dim=-1) + start_logits = start_logits.squeeze(-1) + end_logits = end_logits.squeeze(-1) + + total_loss = None + if start_positions is not None and end_positions is not None: + # If we are on multi-GPU, split add a dimension + if len(start_positions.size()) > 1: + start_positions = start_positions.squeeze(-1) + if len(end_positions.size()) > 1: + end_positions = end_positions.squeeze(-1) + # sometimes the start/end positions are outside our model inputs, we ignore these terms + ignored_index = start_logits.size(1) + start_positions.clamp_(0, ignored_index) + end_positions.clamp_(0, ignored_index) + + loss_fct = CrossEntropyLoss(ignore_index=ignored_index) + start_loss = loss_fct(start_logits, start_positions) + end_loss = loss_fct(end_logits, end_positions) + total_loss = (start_loss + end_loss) / 2 + + if not return_dict: + output = (start_logits, end_logits) + outputs[2:] + return ((total_loss,) + output) if total_loss is not None else output + + return QuestionAnsweringModelOutput( + loss=total_loss, + start_logits=start_logits, + end_logits=end_logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/modeling_tf_rembert.py b/venv/lib/python3.10/site-packages/transformers/models/rembert/modeling_tf_rembert.py new file mode 100644 index 0000000000000000000000000000000000000000..94667c25379b02c060a61c82f04fe451bc22a6be --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/rembert/modeling_tf_rembert.py @@ -0,0 +1,1712 @@ +# coding=utf-8 +# Copyright 2021 The HuggingFace Team and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" TF 2.0 RemBERT model.""" + + +from __future__ import annotations + +import math +from typing import Dict, Optional, Tuple, Union + +import numpy as np +import tensorflow as tf + +from ...activations_tf import get_tf_activation +from ...modeling_tf_outputs import ( + TFBaseModelOutputWithPastAndCrossAttentions, + TFBaseModelOutputWithPoolingAndCrossAttentions, + TFCausalLMOutputWithCrossAttentions, + TFMaskedLMOutput, + TFMultipleChoiceModelOutput, + TFQuestionAnsweringModelOutput, + TFSequenceClassifierOutput, + TFTokenClassifierOutput, +) +from ...modeling_tf_utils import ( + TFCausalLanguageModelingLoss, + TFMaskedLanguageModelingLoss, + TFModelInputType, + TFMultipleChoiceLoss, + TFPreTrainedModel, + TFQuestionAnsweringLoss, + TFSequenceClassificationLoss, + TFTokenClassificationLoss, + get_initializer, + keras, + keras_serializable, + unpack_inputs, +) +from ...tf_utils import check_embeddings_within_bounds, shape_list, stable_softmax +from ...utils import ( + add_code_sample_docstrings, + add_start_docstrings, + add_start_docstrings_to_model_forward, + logging, +) +from .configuration_rembert import RemBertConfig + + +logger = logging.get_logger(__name__) + +_CONFIG_FOR_DOC = "RemBertConfig" + + +from ..deprecated._archive_maps import TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST # noqa: F401, E402 + + +class TFRemBertEmbeddings(keras.layers.Layer): + """Construct the embeddings from word, position and token_type embeddings.""" + + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + + self.config = config + self.input_embedding_size = config.input_embedding_size + self.max_position_embeddings = config.max_position_embeddings + self.initializer_range = config.initializer_range + self.LayerNorm = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="LayerNorm") + self.dropout = keras.layers.Dropout(rate=config.hidden_dropout_prob) + + def build(self, input_shape=None): + with tf.name_scope("word_embeddings"): + self.weight = self.add_weight( + name="weight", + shape=[self.config.vocab_size, self.input_embedding_size], + initializer=get_initializer(self.initializer_range), + ) + + with tf.name_scope("token_type_embeddings"): + self.token_type_embeddings = self.add_weight( + name="embeddings", + shape=[self.config.type_vocab_size, self.input_embedding_size], + initializer=get_initializer(self.initializer_range), + ) + + with tf.name_scope("position_embeddings"): + self.position_embeddings = self.add_weight( + name="embeddings", + shape=[self.max_position_embeddings, self.input_embedding_size], + initializer=get_initializer(self.initializer_range), + ) + + if self.built: + return + self.built = True + if getattr(self, "LayerNorm", None) is not None: + with tf.name_scope(self.LayerNorm.name): + self.LayerNorm.build([None, None, self.config.input_embedding_size]) + + def call( + self, + input_ids: tf.Tensor = None, + position_ids: tf.Tensor = None, + token_type_ids: tf.Tensor = None, + inputs_embeds: tf.Tensor = None, + past_key_values_length=0, + training: bool = False, + ) -> tf.Tensor: + """ + Applies embedding based on inputs tensor. + + Returns: + final_embeddings (`tf.Tensor`): output embedding tensor. + """ + assert not (input_ids is None and inputs_embeds is None) + + if input_ids is not None: + check_embeddings_within_bounds(input_ids, self.config.vocab_size) + inputs_embeds = tf.gather(params=self.weight, indices=input_ids) + + input_shape = shape_list(inputs_embeds)[:-1] + + if token_type_ids is None: + token_type_ids = tf.fill(dims=input_shape, value=0) + + if position_ids is None: + position_ids = tf.expand_dims( + tf.range(start=past_key_values_length, limit=input_shape[1] + past_key_values_length), axis=0 + ) + + position_embeds = tf.gather(params=self.position_embeddings, indices=position_ids) + token_type_embeds = tf.gather(params=self.token_type_embeddings, indices=token_type_ids) + final_embeddings = inputs_embeds + position_embeds + token_type_embeds + final_embeddings = self.LayerNorm(inputs=final_embeddings) + final_embeddings = self.dropout(inputs=final_embeddings, training=training) + + return final_embeddings + + +# Copied from transformers.models.bert.modeling_tf_bert.TFBertSelfAttention with Bert->RemBert +class TFRemBertSelfAttention(keras.layers.Layer): + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + + if config.hidden_size % config.num_attention_heads != 0: + raise ValueError( + f"The hidden size ({config.hidden_size}) is not a multiple of the number " + f"of attention heads ({config.num_attention_heads})" + ) + + self.num_attention_heads = config.num_attention_heads + self.attention_head_size = int(config.hidden_size / config.num_attention_heads) + self.all_head_size = self.num_attention_heads * self.attention_head_size + self.sqrt_att_head_size = math.sqrt(self.attention_head_size) + + self.query = keras.layers.Dense( + units=self.all_head_size, kernel_initializer=get_initializer(config.initializer_range), name="query" + ) + self.key = keras.layers.Dense( + units=self.all_head_size, kernel_initializer=get_initializer(config.initializer_range), name="key" + ) + self.value = keras.layers.Dense( + units=self.all_head_size, kernel_initializer=get_initializer(config.initializer_range), name="value" + ) + self.dropout = keras.layers.Dropout(rate=config.attention_probs_dropout_prob) + + self.is_decoder = config.is_decoder + self.config = config + + def transpose_for_scores(self, tensor: tf.Tensor, batch_size: int) -> tf.Tensor: + # Reshape from [batch_size, seq_length, all_head_size] to [batch_size, seq_length, num_attention_heads, attention_head_size] + tensor = tf.reshape(tensor=tensor, shape=(batch_size, -1, self.num_attention_heads, self.attention_head_size)) + + # Transpose the tensor from [batch_size, seq_length, num_attention_heads, attention_head_size] to [batch_size, num_attention_heads, seq_length, attention_head_size] + return tf.transpose(tensor, perm=[0, 2, 1, 3]) + + def call( + self, + hidden_states: tf.Tensor, + attention_mask: tf.Tensor, + head_mask: tf.Tensor, + encoder_hidden_states: tf.Tensor, + encoder_attention_mask: tf.Tensor, + past_key_value: Tuple[tf.Tensor], + output_attentions: bool, + training: bool = False, + ) -> Tuple[tf.Tensor]: + batch_size = shape_list(hidden_states)[0] + mixed_query_layer = self.query(inputs=hidden_states) + + # If this is instantiated as a cross-attention module, the keys + # and values come from an encoder; the attention mask needs to be + # such that the encoder's padding tokens are not attended to. + is_cross_attention = encoder_hidden_states is not None + + if is_cross_attention and past_key_value is not None: + # reuse k,v, cross_attentions + key_layer = past_key_value[0] + value_layer = past_key_value[1] + attention_mask = encoder_attention_mask + elif is_cross_attention: + key_layer = self.transpose_for_scores(self.key(inputs=encoder_hidden_states), batch_size) + value_layer = self.transpose_for_scores(self.value(inputs=encoder_hidden_states), batch_size) + attention_mask = encoder_attention_mask + elif past_key_value is not None: + key_layer = self.transpose_for_scores(self.key(inputs=hidden_states), batch_size) + value_layer = self.transpose_for_scores(self.value(inputs=hidden_states), batch_size) + key_layer = tf.concat([past_key_value[0], key_layer], axis=2) + value_layer = tf.concat([past_key_value[1], value_layer], axis=2) + else: + key_layer = self.transpose_for_scores(self.key(inputs=hidden_states), batch_size) + value_layer = self.transpose_for_scores(self.value(inputs=hidden_states), batch_size) + + query_layer = self.transpose_for_scores(mixed_query_layer, batch_size) + + if self.is_decoder: + # if cross_attention save Tuple(tf.Tensor, tf.Tensor) of all cross attention key/value_states. + # Further calls to cross_attention layer can then reuse all cross-attention + # key/value_states (first "if" case) + # if uni-directional self-attention (decoder) save Tuple(tf.Tensor, tf.Tensor) of + # all previous decoder key/value_states. Further calls to uni-directional self-attention + # can concat previous decoder key/value_states to current projected key/value_states (third "elif" case) + # if encoder bi-directional self-attention `past_key_value` is always `None` + past_key_value = (key_layer, value_layer) + + # Take the dot product between "query" and "key" to get the raw attention scores. + # (batch size, num_heads, seq_len_q, seq_len_k) + attention_scores = tf.matmul(query_layer, key_layer, transpose_b=True) + dk = tf.cast(self.sqrt_att_head_size, dtype=attention_scores.dtype) + attention_scores = tf.divide(attention_scores, dk) + + if attention_mask is not None: + # Apply the attention mask is (precomputed for all layers in TFRemBertModel call() function) + attention_scores = tf.add(attention_scores, attention_mask) + + # Normalize the attention scores to probabilities. + attention_probs = stable_softmax(logits=attention_scores, axis=-1) + + # This is actually dropping out entire tokens to attend to, which might + # seem a bit unusual, but is taken from the original Transformer paper. + attention_probs = self.dropout(inputs=attention_probs, training=training) + + # Mask heads if we want to + if head_mask is not None: + attention_probs = tf.multiply(attention_probs, head_mask) + + attention_output = tf.matmul(attention_probs, value_layer) + attention_output = tf.transpose(attention_output, perm=[0, 2, 1, 3]) + + # (batch_size, seq_len_q, all_head_size) + attention_output = tf.reshape(tensor=attention_output, shape=(batch_size, -1, self.all_head_size)) + outputs = (attention_output, attention_probs) if output_attentions else (attention_output,) + + if self.is_decoder: + outputs = outputs + (past_key_value,) + return outputs + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "query", None) is not None: + with tf.name_scope(self.query.name): + self.query.build([None, None, self.config.hidden_size]) + if getattr(self, "key", None) is not None: + with tf.name_scope(self.key.name): + self.key.build([None, None, self.config.hidden_size]) + if getattr(self, "value", None) is not None: + with tf.name_scope(self.value.name): + self.value.build([None, None, self.config.hidden_size]) + + +# Copied from transformers.models.bert.modeling_tf_bert.TFBertSelfOutput with Bert->RemBert +class TFRemBertSelfOutput(keras.layers.Layer): + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + + self.dense = keras.layers.Dense( + units=config.hidden_size, kernel_initializer=get_initializer(config.initializer_range), name="dense" + ) + self.LayerNorm = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="LayerNorm") + self.dropout = keras.layers.Dropout(rate=config.hidden_dropout_prob) + self.config = config + + def call(self, hidden_states: tf.Tensor, input_tensor: tf.Tensor, training: bool = False) -> tf.Tensor: + hidden_states = self.dense(inputs=hidden_states) + hidden_states = self.dropout(inputs=hidden_states, training=training) + hidden_states = self.LayerNorm(inputs=hidden_states + input_tensor) + + return hidden_states + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "dense", None) is not None: + with tf.name_scope(self.dense.name): + self.dense.build([None, None, self.config.hidden_size]) + if getattr(self, "LayerNorm", None) is not None: + with tf.name_scope(self.LayerNorm.name): + self.LayerNorm.build([None, None, self.config.hidden_size]) + + +# Copied from transformers.models.bert.modeling_tf_bert.TFBertAttention with Bert->RemBert +class TFRemBertAttention(keras.layers.Layer): + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + + self.self_attention = TFRemBertSelfAttention(config, name="self") + self.dense_output = TFRemBertSelfOutput(config, name="output") + + def prune_heads(self, heads): + raise NotImplementedError + + def call( + self, + input_tensor: tf.Tensor, + attention_mask: tf.Tensor, + head_mask: tf.Tensor, + encoder_hidden_states: tf.Tensor, + encoder_attention_mask: tf.Tensor, + past_key_value: Tuple[tf.Tensor], + output_attentions: bool, + training: bool = False, + ) -> Tuple[tf.Tensor]: + self_outputs = self.self_attention( + hidden_states=input_tensor, + attention_mask=attention_mask, + head_mask=head_mask, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_attention_mask, + past_key_value=past_key_value, + output_attentions=output_attentions, + training=training, + ) + attention_output = self.dense_output( + hidden_states=self_outputs[0], input_tensor=input_tensor, training=training + ) + # add attentions (possibly with past_key_value) if we output them + outputs = (attention_output,) + self_outputs[1:] + + return outputs + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "self_attention", None) is not None: + with tf.name_scope(self.self_attention.name): + self.self_attention.build(None) + if getattr(self, "dense_output", None) is not None: + with tf.name_scope(self.dense_output.name): + self.dense_output.build(None) + + +# Copied from transformers.models.bert.modeling_tf_bert.TFBertIntermediate with Bert->RemBert +class TFRemBertIntermediate(keras.layers.Layer): + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + + self.dense = keras.layers.Dense( + units=config.intermediate_size, kernel_initializer=get_initializer(config.initializer_range), name="dense" + ) + + if isinstance(config.hidden_act, str): + self.intermediate_act_fn = get_tf_activation(config.hidden_act) + else: + self.intermediate_act_fn = config.hidden_act + self.config = config + + def call(self, hidden_states: tf.Tensor) -> tf.Tensor: + hidden_states = self.dense(inputs=hidden_states) + hidden_states = self.intermediate_act_fn(hidden_states) + + return hidden_states + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "dense", None) is not None: + with tf.name_scope(self.dense.name): + self.dense.build([None, None, self.config.hidden_size]) + + +# Copied from transformers.models.bert.modeling_tf_bert.TFBertOutput with Bert->RemBert +class TFRemBertOutput(keras.layers.Layer): + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + + self.dense = keras.layers.Dense( + units=config.hidden_size, kernel_initializer=get_initializer(config.initializer_range), name="dense" + ) + self.LayerNorm = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="LayerNorm") + self.dropout = keras.layers.Dropout(rate=config.hidden_dropout_prob) + self.config = config + + def call(self, hidden_states: tf.Tensor, input_tensor: tf.Tensor, training: bool = False) -> tf.Tensor: + hidden_states = self.dense(inputs=hidden_states) + hidden_states = self.dropout(inputs=hidden_states, training=training) + hidden_states = self.LayerNorm(inputs=hidden_states + input_tensor) + + return hidden_states + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "dense", None) is not None: + with tf.name_scope(self.dense.name): + self.dense.build([None, None, self.config.intermediate_size]) + if getattr(self, "LayerNorm", None) is not None: + with tf.name_scope(self.LayerNorm.name): + self.LayerNorm.build([None, None, self.config.hidden_size]) + + +# Copied from transformers.models.bert.modeling_tf_bert.TFBertLayer with Bert->RemBert +class TFRemBertLayer(keras.layers.Layer): + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + + self.attention = TFRemBertAttention(config, name="attention") + self.is_decoder = config.is_decoder + self.add_cross_attention = config.add_cross_attention + if self.add_cross_attention: + if not self.is_decoder: + raise ValueError(f"{self} should be used as a decoder model if cross attention is added") + self.crossattention = TFRemBertAttention(config, name="crossattention") + self.intermediate = TFRemBertIntermediate(config, name="intermediate") + self.bert_output = TFRemBertOutput(config, name="output") + + def call( + self, + hidden_states: tf.Tensor, + attention_mask: tf.Tensor, + head_mask: tf.Tensor, + encoder_hidden_states: tf.Tensor | None, + encoder_attention_mask: tf.Tensor | None, + past_key_value: Tuple[tf.Tensor] | None, + output_attentions: bool, + training: bool = False, + ) -> Tuple[tf.Tensor]: + # decoder uni-directional self-attention cached key/values tuple is at positions 1,2 + self_attn_past_key_value = past_key_value[:2] if past_key_value is not None else None + self_attention_outputs = self.attention( + input_tensor=hidden_states, + attention_mask=attention_mask, + head_mask=head_mask, + encoder_hidden_states=None, + encoder_attention_mask=None, + past_key_value=self_attn_past_key_value, + output_attentions=output_attentions, + training=training, + ) + attention_output = self_attention_outputs[0] + + # if decoder, the last output is tuple of self-attn cache + if self.is_decoder: + outputs = self_attention_outputs[1:-1] + present_key_value = self_attention_outputs[-1] + else: + outputs = self_attention_outputs[1:] # add self attentions if we output attention weights + + cross_attn_present_key_value = None + if self.is_decoder and encoder_hidden_states is not None: + if not hasattr(self, "crossattention"): + raise ValueError( + f"If `encoder_hidden_states` are passed, {self} has to be instantiated with cross-attention layers" + " by setting `config.add_cross_attention=True`" + ) + + # cross_attn cached key/values tuple is at positions 3,4 of past_key_value tuple + cross_attn_past_key_value = past_key_value[-2:] if past_key_value is not None else None + cross_attention_outputs = self.crossattention( + input_tensor=attention_output, + attention_mask=attention_mask, + head_mask=head_mask, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_attention_mask, + past_key_value=cross_attn_past_key_value, + output_attentions=output_attentions, + training=training, + ) + attention_output = cross_attention_outputs[0] + outputs = outputs + cross_attention_outputs[1:-1] # add cross attentions if we output attention weights + + # add cross-attn cache to positions 3,4 of present_key_value tuple + cross_attn_present_key_value = cross_attention_outputs[-1] + present_key_value = present_key_value + cross_attn_present_key_value + + intermediate_output = self.intermediate(hidden_states=attention_output) + layer_output = self.bert_output( + hidden_states=intermediate_output, input_tensor=attention_output, training=training + ) + outputs = (layer_output,) + outputs # add attentions if we output them + + # if decoder, return the attn key/values as the last output + if self.is_decoder: + outputs = outputs + (present_key_value,) + + return outputs + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "attention", None) is not None: + with tf.name_scope(self.attention.name): + self.attention.build(None) + if getattr(self, "intermediate", None) is not None: + with tf.name_scope(self.intermediate.name): + self.intermediate.build(None) + if getattr(self, "bert_output", None) is not None: + with tf.name_scope(self.bert_output.name): + self.bert_output.build(None) + if getattr(self, "crossattention", None) is not None: + with tf.name_scope(self.crossattention.name): + self.crossattention.build(None) + + +class TFRemBertEncoder(keras.layers.Layer): + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + self.config = config + + self.embedding_hidden_mapping_in = keras.layers.Dense( + units=config.hidden_size, + kernel_initializer=get_initializer(config.initializer_range), + name="embedding_hidden_mapping_in", + ) + self.layer = [TFRemBertLayer(config, name="layer_._{}".format(i)) for i in range(config.num_hidden_layers)] + + def call( + self, + hidden_states: tf.Tensor, + attention_mask: tf.Tensor, + head_mask: tf.Tensor, + encoder_hidden_states: tf.Tensor, + encoder_attention_mask: tf.Tensor, + past_key_values: Tuple[Tuple[tf.Tensor]], + use_cache: bool, + output_attentions: bool, + output_hidden_states: bool, + return_dict: bool, + training: bool = False, + ) -> Union[TFBaseModelOutputWithPastAndCrossAttentions, Tuple[tf.Tensor]]: + hidden_states = self.embedding_hidden_mapping_in(inputs=hidden_states) + all_hidden_states = () if output_hidden_states else None + all_attentions = () if output_attentions else None + all_cross_attentions = () if output_attentions and self.config.add_cross_attention else None + + next_decoder_cache = () if use_cache else None + for i, layer_module in enumerate(self.layer): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + past_key_value = past_key_values[i] if past_key_values is not None else None + + layer_outputs = layer_module( + hidden_states=hidden_states, + attention_mask=attention_mask, + head_mask=head_mask[i], + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_attention_mask, + past_key_value=past_key_value, + output_attentions=output_attentions, + training=training, + ) + hidden_states = layer_outputs[0] + + if use_cache: + next_decoder_cache += (layer_outputs[-1],) + + if output_attentions: + all_attentions = all_attentions + (layer_outputs[1],) + if self.config.add_cross_attention and encoder_hidden_states is not None: + all_cross_attentions = all_cross_attentions + (layer_outputs[2],) + + # Add last layer + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if not return_dict: + return tuple( + v for v in [hidden_states, all_hidden_states, all_attentions, all_cross_attentions] if v is not None + ) + + return TFBaseModelOutputWithPastAndCrossAttentions( + last_hidden_state=hidden_states, + past_key_values=next_decoder_cache, + hidden_states=all_hidden_states, + attentions=all_attentions, + cross_attentions=all_cross_attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "embedding_hidden_mapping_in", None) is not None: + with tf.name_scope(self.embedding_hidden_mapping_in.name): + self.embedding_hidden_mapping_in.build([None, None, self.config.input_embedding_size]) + if getattr(self, "layer", None) is not None: + for layer in self.layer: + with tf.name_scope(layer.name): + layer.build(None) + + +# Copied from transformers.models.bert.modeling_tf_bert.TFBertPooler with Bert->RemBert +class TFRemBertPooler(keras.layers.Layer): + def __init__(self, config: RemBertConfig, **kwargs): + super().__init__(**kwargs) + + self.dense = keras.layers.Dense( + units=config.hidden_size, + kernel_initializer=get_initializer(config.initializer_range), + activation="tanh", + name="dense", + ) + self.config = config + + def call(self, hidden_states: tf.Tensor) -> tf.Tensor: + # We "pool" the model by simply taking the hidden state corresponding + # to the first token. + first_token_tensor = hidden_states[:, 0] + pooled_output = self.dense(inputs=first_token_tensor) + + return pooled_output + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "dense", None) is not None: + with tf.name_scope(self.dense.name): + self.dense.build([None, None, self.config.hidden_size]) + + +class TFRemBertLMPredictionHead(keras.layers.Layer): + def __init__(self, config: RemBertConfig, input_embeddings: keras.layers.Layer, **kwargs): + super().__init__(**kwargs) + + self.config = config + self.initializer_range = config.initializer_range + self.output_embedding_size = config.output_embedding_size + self.dense = keras.layers.Dense( + config.output_embedding_size, kernel_initializer=get_initializer(self.initializer_range), name="dense" + ) + if isinstance(config.hidden_act, str): + self.activation = get_tf_activation(config.hidden_act) + else: + self.activation = config.hidden_act + self.LayerNorm = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="LayerNorm") + + def build(self, input_shape=None): + self.decoder = self.add_weight( + name="decoder/weight", + shape=[self.config.vocab_size, self.output_embedding_size], + initializer=get_initializer(self.initializer_range), + ) + self.decoder_bias = self.add_weight( + shape=(self.config.vocab_size,), initializer="zeros", trainable=True, name="decoder/bias" + ) + + if self.built: + return + self.built = True + if getattr(self, "dense", None) is not None: + with tf.name_scope(self.dense.name): + self.dense.build([None, None, self.config.hidden_size]) + if getattr(self, "LayerNorm", None) is not None: + with tf.name_scope(self.LayerNorm.name): + self.LayerNorm.build([None, self.config.output_embedding_size]) + + def get_output_embeddings(self) -> keras.layers.Layer: + return self + + def set_output_embeddings(self, value): + self.decoder = value + self.decoder.vocab_size = shape_list(value)[0] + + def get_bias(self) -> Dict[str, tf.Variable]: + return {"decoder_bias": self.decoder_bias} + + def set_bias(self, value: tf.Variable): + self.decoder_bias = value["decoder_bias"] + self.config.vocab_size = shape_list(value["decoder_bias"])[0] + + def call(self, hidden_states: tf.Tensor) -> tf.Tensor: + hidden_states = self.dense(inputs=hidden_states) + hidden_states = self.activation(hidden_states) + seq_length = shape_list(tensor=hidden_states)[1] + hidden_states = tf.reshape(tensor=hidden_states, shape=[-1, self.output_embedding_size]) + hidden_states = self.LayerNorm(hidden_states) + hidden_states = tf.matmul(a=hidden_states, b=self.decoder, transpose_b=True) + hidden_states = tf.reshape(tensor=hidden_states, shape=[-1, seq_length, self.config.vocab_size]) + hidden_states = tf.nn.bias_add(value=hidden_states, bias=self.decoder_bias) + return hidden_states + + +# Copied from transformers.models.bert.modeling_tf_bert.TFBertMLMHead with Bert->RemBert +class TFRemBertMLMHead(keras.layers.Layer): + def __init__(self, config: RemBertConfig, input_embeddings: keras.layers.Layer, **kwargs): + super().__init__(**kwargs) + + self.predictions = TFRemBertLMPredictionHead(config, input_embeddings, name="predictions") + + def call(self, sequence_output: tf.Tensor) -> tf.Tensor: + prediction_scores = self.predictions(hidden_states=sequence_output) + + return prediction_scores + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "predictions", None) is not None: + with tf.name_scope(self.predictions.name): + self.predictions.build(None) + + +@keras_serializable +class TFRemBertMainLayer(keras.layers.Layer): + config_class = RemBertConfig + + def __init__(self, config: RemBertConfig, add_pooling_layer: bool = True, **kwargs): + super().__init__(**kwargs) + + self.config = config + self.is_decoder = config.is_decoder + + self.embeddings = TFRemBertEmbeddings(config, name="embeddings") + self.encoder = TFRemBertEncoder(config, name="encoder") + self.pooler = TFRemBertPooler(config, name="pooler") if add_pooling_layer else None + + def get_input_embeddings(self) -> keras.layers.Layer: + return self.embeddings + + def set_input_embeddings(self, value: tf.Variable): + self.embeddings.weight = value + self.embeddings.vocab_size = shape_list(value)[0] + + def _prune_heads(self, heads_to_prune): + """ + Prunes heads of the model. heads_to_prune: dict of {layer_num: list of heads to prune in this layer} See base + class PreTrainedModel + """ + raise NotImplementedError + + @unpack_inputs + # Copied from transformers.models.bert.modeling_tf_bert.TFBertMainLayer.call + def call( + self, + input_ids: TFModelInputType | None = None, + attention_mask: np.ndarray | tf.Tensor | None = None, + token_type_ids: np.ndarray | tf.Tensor | None = None, + position_ids: np.ndarray | tf.Tensor | None = None, + head_mask: np.ndarray | tf.Tensor | None = None, + inputs_embeds: np.ndarray | tf.Tensor | None = None, + encoder_hidden_states: np.ndarray | tf.Tensor | None = None, + encoder_attention_mask: np.ndarray | tf.Tensor | None = None, + past_key_values: Optional[Tuple[Tuple[Union[np.ndarray, tf.Tensor]]]] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + training: bool = False, + ) -> Union[TFBaseModelOutputWithPoolingAndCrossAttentions, Tuple[tf.Tensor]]: + if not self.config.is_decoder: + use_cache = False + + if input_ids is not None and inputs_embeds is not None: + raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time") + elif input_ids is not None: + input_shape = shape_list(input_ids) + elif inputs_embeds is not None: + input_shape = shape_list(inputs_embeds)[:-1] + else: + raise ValueError("You have to specify either input_ids or inputs_embeds") + + batch_size, seq_length = input_shape + + if past_key_values is None: + past_key_values_length = 0 + past_key_values = [None] * len(self.encoder.layer) + else: + past_key_values_length = shape_list(past_key_values[0][0])[-2] + + if attention_mask is None: + attention_mask = tf.fill(dims=(batch_size, seq_length + past_key_values_length), value=1) + + if token_type_ids is None: + token_type_ids = tf.fill(dims=input_shape, value=0) + + embedding_output = self.embeddings( + input_ids=input_ids, + position_ids=position_ids, + token_type_ids=token_type_ids, + inputs_embeds=inputs_embeds, + past_key_values_length=past_key_values_length, + training=training, + ) + + # We create a 3D attention mask from a 2D tensor mask. + # Sizes are [batch_size, 1, 1, to_seq_length] + # So we can broadcast to [batch_size, num_heads, from_seq_length, to_seq_length] + # this attention mask is more simple than the triangular masking of causal attention + # used in OpenAI GPT, we just need to prepare the broadcast dimension here. + attention_mask_shape = shape_list(attention_mask) + + mask_seq_length = seq_length + past_key_values_length + # Copied from `modeling_tf_t5.py` + # Provided a padding mask of dimensions [batch_size, mask_seq_length] + # - if the model is a decoder, apply a causal mask in addition to the padding mask + # - if the model is an encoder, make the mask broadcastable to [batch_size, num_heads, mask_seq_length, mask_seq_length] + if self.is_decoder: + seq_ids = tf.range(mask_seq_length) + causal_mask = tf.less_equal( + tf.tile(seq_ids[None, None, :], (batch_size, mask_seq_length, 1)), + seq_ids[None, :, None], + ) + causal_mask = tf.cast(causal_mask, dtype=attention_mask.dtype) + extended_attention_mask = causal_mask * attention_mask[:, None, :] + attention_mask_shape = shape_list(extended_attention_mask) + extended_attention_mask = tf.reshape( + extended_attention_mask, (attention_mask_shape[0], 1, attention_mask_shape[1], attention_mask_shape[2]) + ) + if past_key_values[0] is not None: + # attention_mask needs to be sliced to the shape `[batch_size, 1, from_seq_length - cached_seq_length, to_seq_length] + extended_attention_mask = extended_attention_mask[:, :, -seq_length:, :] + else: + extended_attention_mask = tf.reshape( + attention_mask, (attention_mask_shape[0], 1, 1, attention_mask_shape[1]) + ) + + # Since attention_mask is 1.0 for positions we want to attend and 0.0 for + # masked positions, this operation will create a tensor which is 0.0 for + # positions we want to attend and -10000.0 for masked positions. + # Since we are adding it to the raw scores before the softmax, this is + # effectively the same as removing these entirely. + extended_attention_mask = tf.cast(extended_attention_mask, dtype=embedding_output.dtype) + one_cst = tf.constant(1.0, dtype=embedding_output.dtype) + ten_thousand_cst = tf.constant(-10000.0, dtype=embedding_output.dtype) + extended_attention_mask = tf.multiply(tf.subtract(one_cst, extended_attention_mask), ten_thousand_cst) + + # Copied from `modeling_tf_t5.py` with -1e9 -> -10000 + if self.is_decoder and encoder_attention_mask is not None: + # If a 2D ou 3D attention mask is provided for the cross-attention + # we need to make broadcastable to [batch_size, num_heads, mask_seq_length, mask_seq_length] + # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] + encoder_attention_mask = tf.cast(encoder_attention_mask, dtype=extended_attention_mask.dtype) + num_dims_encoder_attention_mask = len(shape_list(encoder_attention_mask)) + if num_dims_encoder_attention_mask == 3: + encoder_extended_attention_mask = encoder_attention_mask[:, None, :, :] + if num_dims_encoder_attention_mask == 2: + encoder_extended_attention_mask = encoder_attention_mask[:, None, None, :] + + # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition + # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow/transformer/transformer_layers.py#L270 + # encoder_extended_attention_mask = tf.math.equal(encoder_extended_attention_mask, + # tf.transpose(encoder_extended_attention_mask, perm=(-1, -2))) + + encoder_extended_attention_mask = (1.0 - encoder_extended_attention_mask) * -10000.0 + else: + encoder_extended_attention_mask = None + + # Prepare head mask if needed + # 1.0 in head_mask indicate we keep the head + # attention_probs has shape bsz x n_heads x N x N + # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] + # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] + if head_mask is not None: + raise NotImplementedError + else: + head_mask = [None] * self.config.num_hidden_layers + + encoder_outputs = self.encoder( + hidden_states=embedding_output, + attention_mask=extended_attention_mask, + head_mask=head_mask, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_extended_attention_mask, + past_key_values=past_key_values, + use_cache=use_cache, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + + sequence_output = encoder_outputs[0] + pooled_output = self.pooler(hidden_states=sequence_output) if self.pooler is not None else None + + if not return_dict: + return ( + sequence_output, + pooled_output, + ) + encoder_outputs[1:] + + return TFBaseModelOutputWithPoolingAndCrossAttentions( + last_hidden_state=sequence_output, + pooler_output=pooled_output, + past_key_values=encoder_outputs.past_key_values, + hidden_states=encoder_outputs.hidden_states, + attentions=encoder_outputs.attentions, + cross_attentions=encoder_outputs.cross_attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "embeddings", None) is not None: + with tf.name_scope(self.embeddings.name): + self.embeddings.build(None) + if getattr(self, "encoder", None) is not None: + with tf.name_scope(self.encoder.name): + self.encoder.build(None) + if getattr(self, "pooler", None) is not None: + with tf.name_scope(self.pooler.name): + self.pooler.build(None) + + +class TFRemBertPreTrainedModel(TFPreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = RemBertConfig + base_model_prefix = "rembert" + + +REMBERT_START_DOCSTRING = r""" + + This model inherits from [`TFPreTrainedModel`]. Check the superclass documentation for the generic methods the + library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads + etc.) + + This model is also a [keras.Model](https://www.tensorflow.org/api_docs/python/tf/keras/Model) subclass. Use it + as a regular TF 2.0 Keras Model and refer to the TF 2.0 documentation for all matter related to general usage and + behavior. + + + + TensorFlow models and layers in `transformers` accept two formats as input: + + - having all inputs as keyword arguments (like PyTorch models), or + - having all inputs as a list, tuple or dict in the first positional argument. + + The reason the second format is supported is that Keras methods prefer this format when passing inputs to models + and layers. Because of this support, when using methods like `model.fit()` things should "just work" for you - just + pass your inputs and labels in any format that `model.fit()` supports! If, however, you want to use the second + format outside of Keras methods like `fit()` and `predict()`, such as when creating your own layers or models with + the Keras `Functional` API, there are three possibilities you can use to gather all the input Tensors in the first + positional argument: + + - a single Tensor with `input_ids` only and nothing else: `model(input_ids)` + - a list of varying length with one or several input Tensors IN THE ORDER given in the docstring: + `model([input_ids, attention_mask])` or `model([input_ids, attention_mask, token_type_ids])` + - a dictionary with one or several input Tensors associated to the input names given in the docstring: + `model({"input_ids": input_ids, "token_type_ids": token_type_ids})` + + Note that when creating models and layers with + [subclassing](https://keras.io/guides/making_new_layers_and_models_via_subclassing/) then you don't need to worry + about any of this, as you can just pass inputs like you would to any other Python function! + + + + Args: + config ([`RemBertConfig`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. +""" + +REMBERT_INPUTS_DOCSTRING = r""" + Args: + input_ids (`np.ndarray`, `tf.Tensor`, `List[tf.Tensor]` ``Dict[str, tf.Tensor]` or `Dict[str, np.ndarray]` and each example must have the shape `({0})`): + Indices of input sequence tokens in the vocabulary. + + Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.__call__`] and + [`PreTrainedTokenizer.encode`] for details. + + [What are input IDs?](../glossary#input-ids) + attention_mask (`np.ndarray` or `tf.Tensor` of shape `({0})`, *optional*): + Mask to avoid performing attention on padding token indices. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + [What are attention masks?](../glossary#attention-mask) + token_type_ids (`np.ndarray` or `tf.Tensor` of shape `({0})`, *optional*): + Segment token indices to indicate first and second portions of the inputs. Indices are selected in `[0, + 1]`: + + - 0 corresponds to a *sentence A* token, + - 1 corresponds to a *sentence B* token. + + [What are token type IDs?](../glossary#token-type-ids) + position_ids (`np.ndarray` or `tf.Tensor` of shape `({0})`, *optional*): + Indices of positions of each input sequence tokens in the position embeddings. Selected in the range `[0, + config.max_position_embeddings - 1]`. + + [What are position IDs?](../glossary#position-ids) + head_mask (`np.ndarray` or `tf.Tensor` of shape `(num_heads,)` or `(num_layers, num_heads)`, *optional*): + Mask to nullify selected heads of the self-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + inputs_embeds (`np.ndarray` or `tf.Tensor` of shape `({0}, hidden_size)`, *optional*): + Optionally, instead of passing `input_ids` you can choose to directly pass an embedded representation. This + is useful if you want more control over how to convert `input_ids` indices into associated vectors than the + model's internal embedding lookup matrix. + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. This argument can be used only in eager mode, in graph mode the value in the + config will be used instead. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. This argument can be used only in eager mode, in graph mode the value in the config will be + used instead. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. This argument can be used in + eager mode, in graph mode the value will always be set to True. + training (`bool`, *optional*, defaults to `False``): + Whether or not to use the model in training mode (some modules like dropout modules have different + behaviors between training and evaluation). +""" + + +@add_start_docstrings( + "The bare RemBERT Model transformer outputing raw hidden-states without any specific head on top.", + REMBERT_START_DOCSTRING, +) +class TFRemBertModel(TFRemBertPreTrainedModel): + def __init__(self, config: RemBertConfig, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + self.rembert = TFRemBertMainLayer(config, name="rembert") + + @unpack_inputs + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=TFBaseModelOutputWithPoolingAndCrossAttentions, + config_class=_CONFIG_FOR_DOC, + ) + def call( + self, + input_ids: TFModelInputType | None = None, + attention_mask: np.ndarray | tf.Tensor | None = None, + token_type_ids: np.ndarray | tf.Tensor | None = None, + position_ids: np.ndarray | tf.Tensor | None = None, + head_mask: np.ndarray | tf.Tensor | None = None, + inputs_embeds: np.ndarray | tf.Tensor | None = None, + encoder_hidden_states: np.ndarray | tf.Tensor | None = None, + encoder_attention_mask: np.ndarray | tf.Tensor | None = None, + past_key_values: Optional[Tuple[Tuple[Union[np.ndarray, tf.Tensor]]]] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + training: Optional[bool] = False, + ) -> Union[TFBaseModelOutputWithPoolingAndCrossAttentions, Tuple[tf.Tensor]]: + r""" + encoder_hidden_states (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the encoder. Used in the cross-attention if + the model is configured as a decoder. + encoder_attention_mask (`tf.Tensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on the padding token indices of the encoder input. This mask is used in + the cross-attention if the model is configured as a decoder. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + past_key_values (`Tuple[Tuple[tf.Tensor]]` of length `config.n_layers`) + contains precomputed key and value hidden states of the attention blocks. Can be used to speed up decoding. + If `past_key_values` are used, the user can optionally input only the last `decoder_input_ids` (those that + don't have their past key value states given to this model) of shape `(batch_size, 1)` instead of all + `decoder_input_ids` of shape `(batch_size, sequence_length)`. + use_cache (`bool`, *optional*, defaults to `True`): + If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding (see + `past_key_values`). Set to `False` during training, `True` during generation + """ + outputs = self.rembert( + input_ids=input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_attention_mask, + past_key_values=past_key_values, + use_cache=use_cache, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + + return outputs + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "rembert", None) is not None: + with tf.name_scope(self.rembert.name): + self.rembert.build(None) + + +@add_start_docstrings("""RemBERT Model with a `language modeling` head on top.""", REMBERT_START_DOCSTRING) +class TFRemBertForMaskedLM(TFRemBertPreTrainedModel, TFMaskedLanguageModelingLoss): + def __init__(self, config: RemBertConfig, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + if config.is_decoder: + logger.warning( + "If you want to use `TFRemBertForMaskedLM` make sure `config.is_decoder=False` for " + "bi-directional self-attention." + ) + + self.rembert = TFRemBertMainLayer(config, name="rembert", add_pooling_layer=False) + self.mlm = TFRemBertMLMHead(config, input_embeddings=self.rembert.embeddings, name="mlm___cls") + + def get_lm_head(self) -> keras.layers.Layer: + return self.mlm.predictions + + @unpack_inputs + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=TFMaskedLMOutput, + config_class=_CONFIG_FOR_DOC, + ) + def call( + self, + input_ids: TFModelInputType | None = None, + attention_mask: np.ndarray | tf.Tensor | None = None, + token_type_ids: np.ndarray | tf.Tensor | None = None, + position_ids: np.ndarray | tf.Tensor | None = None, + head_mask: np.ndarray | tf.Tensor | None = None, + inputs_embeds: np.ndarray | tf.Tensor | None = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + labels: np.ndarray | tf.Tensor | None = None, + training: Optional[bool] = False, + ) -> Union[TFMaskedLMOutput, Tuple[tf.Tensor]]: + r""" + labels (`tf.Tensor` or `np.ndarray` of shape `(batch_size, sequence_length)`, *optional*): + Labels for computing the masked language modeling loss. Indices should be in `[-100, 0, ..., + config.vocab_size]` (see `input_ids` docstring) Tokens with indices set to `-100` are ignored (masked), the + loss is only computed for the tokens with labels in `[0, ..., config.vocab_size]` + """ + outputs = self.rembert( + input_ids=input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + sequence_output = outputs[0] + prediction_scores = self.mlm(sequence_output=sequence_output, training=training) + loss = None if labels is None else self.hf_compute_loss(labels=labels, logits=prediction_scores) + + if not return_dict: + output = (prediction_scores,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return TFMaskedLMOutput( + loss=loss, + logits=prediction_scores, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "rembert", None) is not None: + with tf.name_scope(self.rembert.name): + self.rembert.build(None) + if getattr(self, "mlm", None) is not None: + with tf.name_scope(self.mlm.name): + self.mlm.build(None) + + +@add_start_docstrings( + """RemBERT Model with a `language modeling` head on top for CLM fine-tuning.""", REMBERT_START_DOCSTRING +) +class TFRemBertForCausalLM(TFRemBertPreTrainedModel, TFCausalLanguageModelingLoss): + def __init__(self, config: RemBertConfig, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + if not config.is_decoder: + logger.warning("If you want to use `TFRemBertForCausalLM` as a standalone, add `is_decoder=True.`") + + self.rembert = TFRemBertMainLayer(config, name="rembert", add_pooling_layer=False) + self.mlm = TFRemBertMLMHead(config, input_embeddings=self.rembert.embeddings, name="mlm___cls") + + def get_lm_head(self) -> keras.layers.Layer: + return self.mlm.predictions + + # Copied from transformers.models.bert.modeling_tf_bert.TFBertLMHeadModel.prepare_inputs_for_generation + def prepare_inputs_for_generation(self, input_ids, past_key_values=None, attention_mask=None, **model_kwargs): + input_shape = input_ids.shape + # if model is used as a decoder in encoder-decoder model, the decoder attention mask is created on the fly + if attention_mask is None: + attention_mask = tf.ones(input_shape) + + # cut decoder_input_ids if past is used + if past_key_values is not None: + input_ids = input_ids[:, -1:] + + return {"input_ids": input_ids, "attention_mask": attention_mask, "past_key_values": past_key_values} + + @unpack_inputs + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=TFCausalLMOutputWithCrossAttentions, + config_class=_CONFIG_FOR_DOC, + ) + def call( + self, + input_ids: TFModelInputType | None = None, + attention_mask: np.ndarray | tf.Tensor | None = None, + token_type_ids: np.ndarray | tf.Tensor | None = None, + position_ids: np.ndarray | tf.Tensor | None = None, + head_mask: np.ndarray | tf.Tensor | None = None, + inputs_embeds: np.ndarray | tf.Tensor | None = None, + encoder_hidden_states: np.ndarray | tf.Tensor | None = None, + encoder_attention_mask: np.ndarray | tf.Tensor | None = None, + past_key_values: Optional[Tuple[Tuple[Union[np.ndarray, tf.Tensor]]]] = None, + use_cache: Optional[bool] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + labels: np.ndarray | tf.Tensor | None = None, + training: Optional[bool] = False, + ) -> Union[TFCausalLMOutputWithCrossAttentions, Tuple[tf.Tensor]]: + r""" + encoder_hidden_states (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`, *optional*): + Sequence of hidden-states at the output of the last layer of the encoder. Used in the cross-attention if + the model is configured as a decoder. + encoder_attention_mask (`tf.Tensor` of shape `(batch_size, sequence_length)`, *optional*): + Mask to avoid performing attention on the padding token indices of the encoder input. This mask is used in + the cross-attention if the model is configured as a decoder. Mask values selected in `[0, 1]`: + + - 1 for tokens that are **not masked**, + - 0 for tokens that are **masked**. + + past_key_values (`Tuple[Tuple[tf.Tensor]]` of length `config.n_layers`) + contains precomputed key and value hidden states of the attention blocks. Can be used to speed up decoding. + If `past_key_values` are used, the user can optionally input only the last `decoder_input_ids` (those that + don't have their past key value states given to this model) of shape `(batch_size, 1)` instead of all + `decoder_input_ids` of shape `(batch_size, sequence_length)`. + use_cache (`bool`, *optional*, defaults to `True`): + If set to `True`, `past_key_values` key value states are returned and can be used to speed up decoding (see + `past_key_values`). Set to `False` during training, `True` during generation + labels (`tf.Tensor` or `np.ndarray` of shape `(batch_size, sequence_length)`, *optional*): + Labels for computing the cross entropy classification loss. Indices should be in `[0, ..., + config.vocab_size - 1]`. + """ + outputs = self.rembert( + input_ids=input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + encoder_hidden_states=encoder_hidden_states, + encoder_attention_mask=encoder_attention_mask, + past_key_values=past_key_values, + use_cache=use_cache, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + sequence_output = outputs[0] + logits = self.mlm(sequence_output=sequence_output, training=training) + loss = None + + if labels is not None: + # shift labels to the left and cut last logit token + shifted_logits = logits[:, :-1] + labels = labels[:, 1:] + loss = self.hf_compute_loss(labels=labels, logits=shifted_logits) + + if not return_dict: + output = (logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return TFCausalLMOutputWithCrossAttentions( + loss=loss, + logits=logits, + past_key_values=outputs.past_key_values, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + cross_attentions=outputs.cross_attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "rembert", None) is not None: + with tf.name_scope(self.rembert.name): + self.rembert.build(None) + if getattr(self, "mlm", None) is not None: + with tf.name_scope(self.mlm.name): + self.mlm.build(None) + + +@add_start_docstrings( + """ + RemBERT Model transformer with a sequence classification/regression head on top e.g., for GLUE tasks. + """, + REMBERT_START_DOCSTRING, +) +class TFRemBertForSequenceClassification(TFRemBertPreTrainedModel, TFSequenceClassificationLoss): + def __init__(self, config: RemBertConfig, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + self.num_labels = config.num_labels + + self.rembert = TFRemBertMainLayer(config, name="rembert") + self.dropout = keras.layers.Dropout(rate=config.classifier_dropout_prob) + self.classifier = keras.layers.Dense( + units=config.num_labels, + kernel_initializer=get_initializer(config.initializer_range), + name="classifier", + ) + self.config = config + + @unpack_inputs + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=TFSequenceClassifierOutput, + config_class=_CONFIG_FOR_DOC, + ) + def call( + self, + input_ids: TFModelInputType | None = None, + attention_mask: np.ndarray | tf.Tensor | None = None, + token_type_ids: np.ndarray | tf.Tensor | None = None, + position_ids: np.ndarray | tf.Tensor | None = None, + head_mask: np.ndarray | tf.Tensor | None = None, + inputs_embeds: np.ndarray | tf.Tensor | None = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + labels: np.ndarray | tf.Tensor | None = None, + training: Optional[bool] = False, + ) -> Union[TFSequenceClassifierOutput, Tuple[tf.Tensor]]: + r""" + labels (`tf.Tensor` or `np.ndarray` of shape `(batch_size,)`, *optional*): + Labels for computing the sequence classification/regression loss. Indices should be in `[0, ..., + config.num_labels - 1]`. If `config.num_labels == 1` a regression loss is computed (Mean-Square loss), If + `config.num_labels > 1` a classification loss is computed (Cross-Entropy). + """ + outputs = self.rembert( + input_ids=input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + pooled_output = outputs[1] + pooled_output = self.dropout(inputs=pooled_output, training=training) + logits = self.classifier(inputs=pooled_output) + loss = None if labels is None else self.hf_compute_loss(labels=labels, logits=logits) + + if not return_dict: + output = (logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return TFSequenceClassifierOutput( + loss=loss, + logits=logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "rembert", None) is not None: + with tf.name_scope(self.rembert.name): + self.rembert.build(None) + if getattr(self, "classifier", None) is not None: + with tf.name_scope(self.classifier.name): + self.classifier.build([None, None, self.config.hidden_size]) + + +@add_start_docstrings( + """ + RemBERT Model with a multiple choice classification head on top (a linear layer on top of the pooled output and a + softmax) e.g. for RocStories/SWAG tasks. + """, + REMBERT_START_DOCSTRING, +) +class TFRemBertForMultipleChoice(TFRemBertPreTrainedModel, TFMultipleChoiceLoss): + def __init__(self, config: RemBertConfig, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + self.rembert = TFRemBertMainLayer(config, name="rembert") + self.dropout = keras.layers.Dropout(rate=config.classifier_dropout_prob) + self.classifier = keras.layers.Dense( + units=1, kernel_initializer=get_initializer(config.initializer_range), name="classifier" + ) + self.config = config + + @unpack_inputs + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, num_choices, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=TFMultipleChoiceModelOutput, + config_class=_CONFIG_FOR_DOC, + ) + def call( + self, + input_ids: TFModelInputType | None = None, + attention_mask: np.ndarray | tf.Tensor | None = None, + token_type_ids: np.ndarray | tf.Tensor | None = None, + position_ids: np.ndarray | tf.Tensor | None = None, + head_mask: np.ndarray | tf.Tensor | None = None, + inputs_embeds: np.ndarray | tf.Tensor | None = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + labels: np.ndarray | tf.Tensor | None = None, + training: Optional[bool] = False, + ) -> Union[TFMultipleChoiceModelOutput, Tuple[tf.Tensor]]: + r""" + labels (`tf.Tensor` or `np.ndarray` of shape `(batch_size,)`, *optional*): + Labels for computing the multiple choice classification loss. Indices should be in `[0, ..., num_choices]` + where `num_choices` is the size of the second dimension of the input tensors. (See `input_ids` above) + """ + + if input_ids is not None: + num_choices = shape_list(input_ids)[1] + seq_length = shape_list(input_ids)[2] + else: + num_choices = shape_list(inputs_embeds)[1] + seq_length = shape_list(inputs_embeds)[2] + + flat_input_ids = tf.reshape(tensor=input_ids, shape=(-1, seq_length)) if input_ids is not None else None + flat_attention_mask = ( + tf.reshape(tensor=attention_mask, shape=(-1, seq_length)) if attention_mask is not None else None + ) + flat_token_type_ids = ( + tf.reshape(tensor=token_type_ids, shape=(-1, seq_length)) if token_type_ids is not None else None + ) + flat_position_ids = ( + tf.reshape(tensor=position_ids, shape=(-1, seq_length)) if position_ids is not None else None + ) + flat_inputs_embeds = ( + tf.reshape(tensor=inputs_embeds, shape=(-1, seq_length, shape_list(inputs_embeds)[3])) + if inputs_embeds is not None + else None + ) + outputs = self.rembert( + input_ids=flat_input_ids, + attention_mask=flat_attention_mask, + token_type_ids=flat_token_type_ids, + position_ids=flat_position_ids, + head_mask=head_mask, + inputs_embeds=flat_inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + pooled_output = outputs[1] + pooled_output = self.dropout(inputs=pooled_output, training=training) + logits = self.classifier(inputs=pooled_output) + reshaped_logits = tf.reshape(tensor=logits, shape=(-1, num_choices)) + loss = None if labels is None else self.hf_compute_loss(labels=labels, logits=reshaped_logits) + + if not return_dict: + output = (reshaped_logits,) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return TFMultipleChoiceModelOutput( + loss=loss, + logits=reshaped_logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "rembert", None) is not None: + with tf.name_scope(self.rembert.name): + self.rembert.build(None) + if getattr(self, "classifier", None) is not None: + with tf.name_scope(self.classifier.name): + self.classifier.build([None, None, self.config.hidden_size]) + + +@add_start_docstrings( + """ + RemBERT Model with a token classification head on top (a linear layer on top of the hidden-states output) e.g. for + Named-Entity-Recognition (NER) tasks. + """, + REMBERT_START_DOCSTRING, +) +class TFRemBertForTokenClassification(TFRemBertPreTrainedModel, TFTokenClassificationLoss): + def __init__(self, config: RemBertConfig, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + self.num_labels = config.num_labels + + self.rembert = TFRemBertMainLayer(config, name="rembert", add_pooling_layer=False) + self.dropout = keras.layers.Dropout(rate=config.hidden_dropout_prob) + self.classifier = keras.layers.Dense( + units=config.num_labels, kernel_initializer=get_initializer(config.initializer_range), name="classifier" + ) + self.config = config + + @unpack_inputs + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=TFTokenClassifierOutput, + config_class=_CONFIG_FOR_DOC, + ) + def call( + self, + input_ids: TFModelInputType | None = None, + attention_mask: np.ndarray | tf.Tensor | None = None, + token_type_ids: np.ndarray | tf.Tensor | None = None, + position_ids: np.ndarray | tf.Tensor | None = None, + head_mask: np.ndarray | tf.Tensor | None = None, + inputs_embeds: np.ndarray | tf.Tensor | None = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + labels: np.ndarray | tf.Tensor | None = None, + training: Optional[bool] = False, + ) -> Union[TFTokenClassifierOutput, Tuple[tf.Tensor]]: + r""" + labels (`tf.Tensor` or `np.ndarray` of shape `(batch_size, sequence_length)`, *optional*): + Labels for computing the token classification loss. Indices should be in `[0, ..., config.num_labels - 1]`. + """ + outputs = self.rembert( + input_ids=input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + sequence_output = outputs[0] + sequence_output = self.dropout(inputs=sequence_output, training=training) + logits = self.classifier(inputs=sequence_output) + loss = None if labels is None else self.hf_compute_loss(labels=labels, logits=logits) + + if not return_dict: + output = (logits,) + outputs[1:] + return ((loss,) + output) if loss is not None else output + + return TFTokenClassifierOutput( + loss=loss, + logits=logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "rembert", None) is not None: + with tf.name_scope(self.rembert.name): + self.rembert.build(None) + if getattr(self, "classifier", None) is not None: + with tf.name_scope(self.classifier.name): + self.classifier.build([None, None, self.config.hidden_size]) + + +@add_start_docstrings( + """ + RemBERT Model with a span classification head on top for extractive question-answering tasks like SQuAD (a linear + layer on top of the hidden-states output to compute `span start logits` and `span end logits`). + """, + REMBERT_START_DOCSTRING, +) +class TFRemBertForQuestionAnswering(TFRemBertPreTrainedModel, TFQuestionAnsweringLoss): + def __init__(self, config: RemBertConfig, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + self.num_labels = config.num_labels + + self.rembert = TFRemBertMainLayer(config, add_pooling_layer=False, name="rembert") + self.qa_outputs = keras.layers.Dense( + units=config.num_labels, kernel_initializer=get_initializer(config.initializer_range), name="qa_outputs" + ) + self.config = config + + @unpack_inputs + @add_start_docstrings_to_model_forward(REMBERT_INPUTS_DOCSTRING.format("batch_size, sequence_length")) + @add_code_sample_docstrings( + checkpoint="google/rembert", + output_type=TFQuestionAnsweringModelOutput, + config_class=_CONFIG_FOR_DOC, + ) + def call( + self, + input_ids: TFModelInputType | None = None, + attention_mask: np.ndarray | tf.Tensor | None = None, + token_type_ids: np.ndarray | tf.Tensor | None = None, + position_ids: np.ndarray | tf.Tensor | None = None, + head_mask: np.ndarray | tf.Tensor | None = None, + inputs_embeds: np.ndarray | tf.Tensor | None = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + start_positions: np.ndarray | tf.Tensor | None = None, + end_positions: np.ndarray | tf.Tensor | None = None, + training: Optional[bool] = False, + ) -> Union[TFQuestionAnsweringModelOutput, Tuple[tf.Tensor]]: + r""" + start_positions (`tf.Tensor` or `np.ndarray` of shape `(batch_size,)`, *optional*): + Labels for position (index) of the start of the labelled span for computing the token classification loss. + Positions are clamped to the length of the sequence (`sequence_length`). Position outside of the sequence + are not taken into account for computing the loss. + end_positions (`tf.Tensor` or `np.ndarray` of shape `(batch_size,)`, *optional*): + Labels for position (index) of the end of the labelled span for computing the token classification loss. + Positions are clamped to the length of the sequence (`sequence_length`). Position outside of the sequence + are not taken into account for computing the loss. + """ + outputs = self.rembert( + input_ids=input_ids, + attention_mask=attention_mask, + token_type_ids=token_type_ids, + position_ids=position_ids, + head_mask=head_mask, + inputs_embeds=inputs_embeds, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + sequence_output = outputs[0] + logits = self.qa_outputs(inputs=sequence_output) + start_logits, end_logits = tf.split(value=logits, num_or_size_splits=2, axis=-1) + start_logits = tf.squeeze(input=start_logits, axis=-1) + end_logits = tf.squeeze(input=end_logits, axis=-1) + loss = None + + if start_positions is not None and end_positions is not None: + labels = {"start_position": start_positions} + labels["end_position"] = end_positions + loss = self.hf_compute_loss(labels=labels, logits=(start_logits, end_logits)) + + if not return_dict: + output = (start_logits, end_logits) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return TFQuestionAnsweringModelOutput( + loss=loss, + start_logits=start_logits, + end_logits=end_logits, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "rembert", None) is not None: + with tf.name_scope(self.rembert.name): + self.rembert.build(None) + if getattr(self, "qa_outputs", None) is not None: + with tf.name_scope(self.qa_outputs.name): + self.qa_outputs.build([None, None, self.config.hidden_size]) diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/tokenization_rembert.py b/venv/lib/python3.10/site-packages/transformers/models/rembert/tokenization_rembert.py new file mode 100644 index 0000000000000000000000000000000000000000..a2b1f9abc2c989100dded2c137bb6d9e3b8f682d --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/rembert/tokenization_rembert.py @@ -0,0 +1,263 @@ +# coding=utf-8 +# Copyright The HuggingFace Team and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Tokenization classes for RemBERT.""" + + +import os +from shutil import copyfile +from typing import List, Optional, Tuple + +import sentencepiece as spm + +from ...tokenization_utils import AddedToken, PreTrainedTokenizer +from ...utils import logging + + +logger = logging.get_logger(__name__) + +VOCAB_FILES_NAMES = {"vocab_file": "sentencepiece.model"} + + +class RemBertTokenizer(PreTrainedTokenizer): + """ + Construct a RemBERT tokenizer. Based on [SentencePiece](https://github.com/google/sentencepiece). + + This tokenizer inherits from [`PreTrainedTokenizer`] which contains most of the main methods. Users should refer to + this superclass for more information regarding those methods. + + Args: + vocab_file (`str`): + [SentencePiece](https://github.com/google/sentencepiece) file (generally has a *.spm* extension) that + contains the vocabulary necessary to instantiate a tokenizer. + bos_token (`str`, *optional*, defaults to `"[CLS]"`): + The beginning of sequence token that was used during pretraining. Can be used a sequence classifier token. + + + + When building a sequence using special tokens, this is not the token that is used for the beginning of + sequence. The token used is the `cls_token`. + + + + eos_token (`str`, *optional*, defaults to `"[SEP]"`): + The end of sequence token. + + + + When building a sequence using special tokens, this is not the token that is used for the end of sequence. + The token used is the `sep_token`. + + + + unk_token (`str`, *optional*, defaults to `""`): + The unknown token. A token that is not in the vocabulary cannot be converted to an ID and is set to be this + token instead. + sep_token (`str`, *optional*, defaults to `"[SEP]"`): + The separator token, which is used when building a sequence from multiple sequences, e.g. two sequences for + sequence classification or for a text and a question for question answering. It is also used as the last + token of a sequence built with special tokens. + pad_token (`str`, *optional*, defaults to `""`): + The token used for padding, for example when batching sequences of different lengths. + cls_token (`str`, *optional*, defaults to `"[CLS]"`): + The classifier token which is used when doing sequence classification (classification of the whole sequence + instead of per-token classification). It is the first token of the sequence when built with special tokens. + mask_token (`str`, *optional*, defaults to `"[MASK]"`): + The token used for masking values. This is the token used when training this model with masked language + modeling. This is the token which the model will try to predict. + + Attributes: + sp_model (`SentencePieceProcessor`): + The *SentencePiece* processor that is used for every conversion (string, tokens and IDs). + """ + + vocab_files_names = VOCAB_FILES_NAMES + + def __init__( + self, + vocab_file, + do_lower_case=False, + remove_space=True, + keep_accents=True, + bos_token="[CLS]", + eos_token="[SEP]", + unk_token="[UNK]", + sep_token="[SEP]", + pad_token="[PAD]", + cls_token="[CLS]", + mask_token="[MASK]", + **kwargs, + ): + # Mask token behave like a normal word, i.e. include the space before it + mask_token = AddedToken(mask_token, lstrip=True, rstrip=False) if isinstance(mask_token, str) else mask_token + + self.do_lower_case = do_lower_case + self.remove_space = remove_space + self.keep_accents = keep_accents + self.vocab_file = vocab_file + + self.sp_model = spm.SentencePieceProcessor() + self.sp_model.Load(vocab_file) + super().__init__( + do_lower_case=do_lower_case, + remove_space=remove_space, + keep_accents=keep_accents, + bos_token=bos_token, + eos_token=eos_token, + unk_token=unk_token, + sep_token=sep_token, + pad_token=pad_token, + cls_token=cls_token, + mask_token=mask_token, + **kwargs, + ) + + @property + def vocab_size(self): + return len(self.sp_model) + + def get_vocab(self): + vocab = {self.convert_ids_to_tokens(i): i for i in range(self.vocab_size)} + vocab.update(self.added_tokens_encoder) + return vocab + + def __getstate__(self): + state = self.__dict__.copy() + state["sp_model"] = None + return state + + def __setstate__(self, d): + self.__dict__ = d + self.sp_model = spm.SentencePieceProcessor() + self.sp_model.Load(self.vocab_file) + + def _tokenize(self, text, sample=False): + """Tokenize a string.""" + pieces = self.sp_model.EncodeAsPieces(text) + return pieces + + def _convert_token_to_id(self, token): + """Converts a token (str) in an id using the vocab.""" + return self.sp_model.PieceToId(token) + + def _convert_id_to_token(self, index): + """Converts an index (integer) in a token (str) using the vocab.""" + return self.sp_model.IdToPiece(index) + + def convert_tokens_to_string(self, tokens): + out_string = self.sp_model.decode_pieces(tokens) + return out_string + + def build_inputs_with_special_tokens( + self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None + ) -> List[int]: + """ + Build model inputs from a sequence or a pair of sequence for sequence classification tasks by concatenating and + adding special tokens. A REMBERT sequence has the following format: + + - single sequence: `[CLS] X [SEP]` + - pair of sequences: `[CLS] A [SEP] B [SEP]` + + Args: + token_ids_0 (`List[int]`): + List of IDs to which the special tokens will be added. + token_ids_1 (`List[int]`, *optional*): + Optional second list of IDs for sequence pairs. + + Returns: + `List[int]`: List of [input IDs](../glossary#input-ids) with the appropriate special tokens. + """ + sep = [self.sep_token_id] + cls = [self.cls_token_id] + if token_ids_1 is None: + return cls + token_ids_0 + sep + return cls + token_ids_0 + sep + token_ids_1 + sep + + def get_special_tokens_mask( + self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None, already_has_special_tokens: bool = False + ) -> List[int]: + """ + Retrieve sequence ids from a token list that has no special tokens added. This method is called when adding + special tokens using the tokenizer `prepare_for_model` method. + + Args: + token_ids_0 (`List[int]`): + List of IDs. + token_ids_1 (`List[int]`, *optional*): + Optional second list of IDs for sequence pairs. + already_has_special_tokens (`bool`, *optional*, defaults to `False`): + Whether or not the token list is already formatted with special tokens for the model. + + Returns: + `List[int]`: A list of integers in the range [0, 1]: 1 for a special token, 0 for a sequence token. + """ + + if already_has_special_tokens: + if token_ids_1 is not None: + raise ValueError( + "You should not supply a second sequence if the provided sequence of " + "ids is already formatted with special tokens for the model." + ) + return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_0] + + if token_ids_1 is not None: + return [1] + ([0] * len(token_ids_0)) + [1] + ([0] * len(token_ids_1)) + [1] + return [1] + ([0] * len(token_ids_0)) + [1] + + def create_token_type_ids_from_sequences( + self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None + ) -> List[int]: + """ + Create a mask from the two sequences passed to be used in a sequence-pair classification task. A RemBERT + sequence pair mask has the following format: + + ``` + 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 + | first sequence | second sequence | + ``` + + If `token_ids_1` is `None`, this method only returns the first portion of the mask (0s). + + Args: + token_ids_0 (`List[int]`): + List of IDs. + token_ids_1 (`List[int]`, *optional*): + Optional second list of IDs for sequence pairs. + + Returns: + `List[int]`: List of [token type IDs](../glossary#token-type-ids) according to the given sequence(s). + """ + sep = [self.sep_token_id] + cls = [self.cls_token_id] + + if token_ids_1 is None: + return len(cls + token_ids_0 + sep) * [0] + return len(cls + token_ids_0 + sep) * [0] + len(token_ids_1 + sep) * [1] + + def save_vocabulary(self, save_directory: str, filename_prefix: Optional[str] = None) -> Tuple[str]: + if not os.path.isdir(save_directory): + logger.error("Vocabulary path ({}) should be a directory".format(save_directory)) + return + out_vocab_file = os.path.join( + save_directory, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] + ) + + if os.path.abspath(self.vocab_file) != os.path.abspath(out_vocab_file) and os.path.isfile(self.vocab_file): + copyfile(self.vocab_file, out_vocab_file) + elif not os.path.isfile(self.vocab_file): + with open(out_vocab_file, "wb") as fi: + content_spiece_model = self.sp_model.serialized_model_proto() + fi.write(content_spiece_model) + + return (out_vocab_file,) diff --git a/venv/lib/python3.10/site-packages/transformers/models/rembert/tokenization_rembert_fast.py b/venv/lib/python3.10/site-packages/transformers/models/rembert/tokenization_rembert_fast.py new file mode 100644 index 0000000000000000000000000000000000000000..b7165e362a4f7aada4c2fd2d65ac089107e14161 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/rembert/tokenization_rembert_fast.py @@ -0,0 +1,230 @@ +# coding=utf-8 +# Copyright 2018 Google AI, Google Brain and the HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" Tokenization classes for RemBERT model.""" + + +import os +from shutil import copyfile +from typing import List, Optional, Tuple + +from ...tokenization_utils import AddedToken +from ...tokenization_utils_fast import PreTrainedTokenizerFast +from ...utils import is_sentencepiece_available, logging + + +if is_sentencepiece_available(): + from .tokenization_rembert import RemBertTokenizer +else: + RemBertTokenizer = None + +logger = logging.get_logger(__name__) +VOCAB_FILES_NAMES = {"vocab_file": "sentencepiece.model", "tokenizer_file": "tokenizer.json"} + + +SPIECE_UNDERLINE = "▁" + + +class RemBertTokenizerFast(PreTrainedTokenizerFast): + """ + Construct a "fast" RemBert tokenizer (backed by HuggingFace's *tokenizers* library). Based on + [Unigram](https://huggingface.co/docs/tokenizers/python/latest/components.html?highlight=unigram#models). This + tokenizer inherits from [`PreTrainedTokenizerFast`] which contains most of the main methods. Users should refer to + this superclass for more information regarding those methods + + Args: + vocab_file (`str`): + [SentencePiece](https://github.com/google/sentencepiece) file (generally has a *.spm* extension) that + contains the vocabulary necessary to instantiate a tokenizer. + do_lower_case (`bool`, *optional*, defaults to `True`): + Whether or not to lowercase the input when tokenizing. + remove_space (`bool`, *optional*, defaults to `True`): + Whether or not to strip the text when tokenizing (removing excess spaces before and after the string). + keep_accents (`bool`, *optional*, defaults to `False`): + Whether or not to keep accents when tokenizing. + bos_token (`str`, *optional*, defaults to `"[CLS]"`): + The beginning of sequence token that was used during pretraining. Can be used a sequence classifier token. + + + + When building a sequence using special tokens, this is not the token that is used for the beginning of + sequence. The token used is the `cls_token`. + + + + eos_token (`str`, *optional*, defaults to `"[SEP]"`): + The end of sequence token. .. note:: When building a sequence using special tokens, this is not the token + that is used for the end of sequence. The token used is the `sep_token`. + unk_token (`str`, *optional*, defaults to `""`): + The unknown token. A token that is not in the vocabulary cannot be converted to an ID and is set to be this + token instead. + sep_token (`str`, *optional*, defaults to `"[SEP]"`): + The separator token, which is used when building a sequence from multiple sequences, e.g. two sequences for + sequence classification or for a text and a question for question answering. It is also used as the last + token of a sequence built with special tokens. + pad_token (`str`, *optional*, defaults to `""`): + The token used for padding, for example when batching sequences of different lengths. + cls_token (`str`, *optional*, defaults to `"[CLS]"`): + The classifier token which is used when doing sequence classification (classification of the whole sequence + instead of per-token classification). It is the first token of the sequence when built with special tokens. + mask_token (`str`, *optional*, defaults to `"[MASK]"`): + The token used for masking values. This is the token used when training this model with masked language + modeling. This is the token which the model will try to predict. + """ + + vocab_files_names = VOCAB_FILES_NAMES + slow_tokenizer_class = RemBertTokenizer + + def __init__( + self, + vocab_file=None, + tokenizer_file=None, + do_lower_case=True, + remove_space=True, + keep_accents=False, + bos_token="[CLS]", + eos_token="[SEP]", + unk_token="", + sep_token="[SEP]", + pad_token="", + cls_token="[CLS]", + mask_token="[MASK]", + **kwargs, + ): + # Mask token behave like a normal word, i.e. include the space before it + mask_token = AddedToken(mask_token, lstrip=True, rstrip=False) if isinstance(mask_token, str) else mask_token + + super().__init__( + vocab_file, + tokenizer_file=tokenizer_file, + do_lower_case=do_lower_case, + remove_space=remove_space, + keep_accents=keep_accents, + bos_token=bos_token, + eos_token=eos_token, + unk_token=unk_token, + sep_token=sep_token, + pad_token=pad_token, + cls_token=cls_token, + mask_token=mask_token, + **kwargs, + ) + + self.do_lower_case = do_lower_case + self.remove_space = remove_space + self.keep_accents = keep_accents + self.vocab_file = vocab_file + + @property + def can_save_slow_tokenizer(self) -> bool: + return os.path.isfile(self.vocab_file) if self.vocab_file else False + + def build_inputs_with_special_tokens( + self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None + ) -> List[int]: + """ + Build model inputs from a sequence or a pair of sequence for sequence classification tasks by concatenating and + adding special tokens. A RemBERT sequence has the following format: + + - single sequence: `[CLS] X [SEP]` + - pair of sequences: `[CLS] A [SEP] B [SEP]` + + Args: + token_ids_0 (`List[int]`): + List of IDs to which the special tokens will be added + token_ids_1 (`List[int]`, *optional*, defaults to `None`): + Optional second list of IDs for sequence pairs. + + Returns: + `List[int]`: list of [input IDs](../glossary#input-ids) with the appropriate special tokens. + """ + sep = [self.sep_token_id] + cls = [self.cls_token_id] + if token_ids_1 is None: + return cls + token_ids_0 + sep + return cls + token_ids_0 + sep + token_ids_1 + sep + + def get_special_tokens_mask( + self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None, already_has_special_tokens: bool = False + ) -> List[int]: + """ + Retrieves sequence ids from a token list that has no special tokens added. This method is called when adding + special tokens using the tokenizer `prepare_for_model` method. + + Args: + token_ids_0 (`List[int]`): + List of ids. + token_ids_1 (`List[int]`, *optional*, defaults to `None`): + Optional second list of IDs for sequence pairs. + already_has_special_tokens (`bool`, *optional*, defaults to `False`): + Set to True if the token list is already formatted with special tokens for the model + + Returns: + `List[int]`: A list of integers in the range [0, 1]: 1 for a special token, 0 for a sequence token. + """ + + if already_has_special_tokens: + if token_ids_1 is not None: + raise ValueError( + "You should not supply a second sequence if the provided sequence of " + "ids is already formatted with special tokens for the model." + ) + return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_0] + + if token_ids_1 is not None: + return [1] + ([0] * len(token_ids_0)) + [1] + ([0] * len(token_ids_1)) + [1] + return [1] + ([0] * len(token_ids_0)) + [1] + + def create_token_type_ids_from_sequences( + self, token_ids_0: List[int], token_ids_1: Optional[List[int]] = None + ) -> List[int]: + """ + Creates a mask from the two sequences passed to be used in a sequence-pair classification task. A RemBERT + sequence pair mask has the following format: + + ``` + 0 0 0 0 0 0 0 0 0 0 0 1 1 1 1 1 1 1 1 1 + | first sequence | second sequence | + ``` + + if token_ids_1 is None, only returns the first portion of the mask (0s). + + Args: + token_ids_0 (`List[int]`): + List of ids. + token_ids_1 (`List[int]`, *optional*, defaults to `None`): + Optional second list of IDs for sequence pairs. + + Returns: + `List[int]`: List of [token type IDs](../glossary#token-type-ids) according to the given sequence(s). + """ + sep = [self.sep_token_id] + cls = [self.cls_token_id] + + if token_ids_1 is None: + return len(cls + token_ids_0 + sep) * [0] + return len(cls + token_ids_0 + sep) * [0] + len(token_ids_1 + sep) * [1] + + def save_vocabulary(self, save_directory: str, filename_prefix: Optional[str] = None) -> Tuple[str]: + if not os.path.isdir(save_directory): + logger.error("Vocabulary path ({}) should be a directory".format(save_directory)) + return + out_vocab_file = os.path.join( + save_directory, (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] + ) + + if os.path.abspath(self.vocab_file) != os.path.abspath(out_vocab_file): + copyfile(self.vocab_file, out_vocab_file) + + return (out_vocab_file,) diff --git a/venv/lib/python3.10/site-packages/transformers/models/vit_mae/__init__.py b/venv/lib/python3.10/site-packages/transformers/models/vit_mae/__init__.py new file mode 100644 index 0000000000000000000000000000000000000000..bfd200e9dcb9130c8612b97295d0721403271555 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/vit_mae/__init__.py @@ -0,0 +1,82 @@ +# Copyright 2020 The HuggingFace Team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +from typing import TYPE_CHECKING + +from ...utils import ( + OptionalDependencyNotAvailable, + _LazyModule, + is_flax_available, + is_tf_available, + is_torch_available, +) + + +_import_structure = {"configuration_vit_mae": ["VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP", "ViTMAEConfig"]} + +try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["modeling_vit_mae"] = [ + "VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST", + "ViTMAEForPreTraining", + "ViTMAELayer", + "ViTMAEModel", + "ViTMAEPreTrainedModel", + ] + +try: + if not is_tf_available(): + raise OptionalDependencyNotAvailable() +except OptionalDependencyNotAvailable: + pass +else: + _import_structure["modeling_tf_vit_mae"] = [ + "TFViTMAEForPreTraining", + "TFViTMAEModel", + "TFViTMAEPreTrainedModel", + ] + +if TYPE_CHECKING: + from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig + + try: + if not is_torch_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .modeling_vit_mae import ( + VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, + ViTMAEForPreTraining, + ViTMAELayer, + ViTMAEModel, + ViTMAEPreTrainedModel, + ) + + try: + if not is_tf_available(): + raise OptionalDependencyNotAvailable() + except OptionalDependencyNotAvailable: + pass + else: + from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel + + +else: + import sys + + sys.modules[__name__] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__) diff --git a/venv/lib/python3.10/site-packages/transformers/models/vit_mae/convert_vit_mae_to_pytorch.py b/venv/lib/python3.10/site-packages/transformers/models/vit_mae/convert_vit_mae_to_pytorch.py new file mode 100644 index 0000000000000000000000000000000000000000..47e77593f6fd3ad7c2b7ff2c329b84f432060c7d --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/vit_mae/convert_vit_mae_to_pytorch.py @@ -0,0 +1,178 @@ +# coding=utf-8 +# Copyright 2022 The HuggingFace Inc. team. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +"""Convert ViT MAE checkpoints from the original repository: https://github.com/facebookresearch/mae""" + +import argparse + +import requests +import torch +from PIL import Image + +from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor + + +def rename_key(name): + if "cls_token" in name: + name = name.replace("cls_token", "vit.embeddings.cls_token") + if "mask_token" in name: + name = name.replace("mask_token", "decoder.mask_token") + if "decoder_pos_embed" in name: + name = name.replace("decoder_pos_embed", "decoder.decoder_pos_embed") + if "pos_embed" in name and "decoder" not in name: + name = name.replace("pos_embed", "vit.embeddings.position_embeddings") + if "patch_embed.proj" in name: + name = name.replace("patch_embed.proj", "vit.embeddings.patch_embeddings.projection") + if "patch_embed.norm" in name: + name = name.replace("patch_embed.norm", "vit.embeddings.norm") + if "decoder_blocks" in name: + name = name.replace("decoder_blocks", "decoder.decoder_layers") + if "blocks" in name: + name = name.replace("blocks", "vit.encoder.layer") + if "attn.proj" in name: + name = name.replace("attn.proj", "attention.output.dense") + if "attn" in name: + name = name.replace("attn", "attention.self") + if "norm1" in name: + name = name.replace("norm1", "layernorm_before") + if "norm2" in name: + name = name.replace("norm2", "layernorm_after") + if "mlp.fc1" in name: + name = name.replace("mlp.fc1", "intermediate.dense") + if "mlp.fc2" in name: + name = name.replace("mlp.fc2", "output.dense") + if "decoder_embed" in name: + name = name.replace("decoder_embed", "decoder.decoder_embed") + if "decoder_norm" in name: + name = name.replace("decoder_norm", "decoder.decoder_norm") + if "decoder_pred" in name: + name = name.replace("decoder_pred", "decoder.decoder_pred") + if "norm.weight" in name and "decoder" not in name: + name = name.replace("norm.weight", "vit.layernorm.weight") + if "norm.bias" in name and "decoder" not in name: + name = name.replace("norm.bias", "vit.layernorm.bias") + + return name + + +def convert_state_dict(orig_state_dict, config): + for key in orig_state_dict.copy().keys(): + val = orig_state_dict.pop(key) + + if "qkv" in key: + key_split = key.split(".") + layer_num = int(key_split[1]) + if "decoder_blocks" in key: + dim = config.decoder_hidden_size + prefix = "decoder.decoder_layers." + if "weight" in key: + orig_state_dict[f"{prefix}{layer_num}.attention.attention.query.weight"] = val[:dim, :] + orig_state_dict[f"{prefix}{layer_num}.attention.attention.key.weight"] = val[dim : dim * 2, :] + orig_state_dict[f"{prefix}{layer_num}.attention.attention.value.weight"] = val[-dim:, :] + elif "bias" in key: + orig_state_dict[f"{prefix}{layer_num}.attention.attention.query.bias"] = val[:dim] + orig_state_dict[f"{prefix}{layer_num}.attention.attention.key.bias"] = val[dim : dim * 2] + orig_state_dict[f"{prefix}{layer_num}.attention.attention.value.bias"] = val[-dim:] + else: + dim = config.hidden_size + prefix = "vit.encoder.layer." + if "weight" in key: + orig_state_dict[f"{prefix}{layer_num}.attention.attention.query.weight"] = val[:dim, :] + orig_state_dict[f"{prefix}{layer_num}.attention.attention.key.weight"] = val[dim : dim * 2, :] + orig_state_dict[f"{prefix}{layer_num}.attention.attention.value.weight"] = val[-dim:, :] + elif "bias" in key: + orig_state_dict[f"{prefix}{layer_num}.attention.attention.query.bias"] = val[:dim] + orig_state_dict[f"{prefix}{layer_num}.attention.attention.key.bias"] = val[dim : dim * 2] + orig_state_dict[f"{prefix}{layer_num}.attention.attention.value.bias"] = val[-dim:] + + else: + orig_state_dict[rename_key(key)] = val + + return orig_state_dict + + +def convert_vit_mae_checkpoint(checkpoint_url, pytorch_dump_folder_path): + config = ViTMAEConfig() + if "large" in checkpoint_url: + config.hidden_size = 1024 + config.intermediate_size = 4096 + config.num_hidden_layers = 24 + config.num_attention_heads = 16 + elif "huge" in checkpoint_url: + config.patch_size = 14 + config.hidden_size = 1280 + config.intermediate_size = 5120 + config.num_hidden_layers = 32 + config.num_attention_heads = 16 + + model = ViTMAEForPreTraining(config) + + state_dict = torch.hub.load_state_dict_from_url(checkpoint_url, map_location="cpu")["model"] + + image_processor = ViTMAEImageProcessor(size=config.image_size) + + new_state_dict = convert_state_dict(state_dict, config) + + model.load_state_dict(new_state_dict) + model.eval() + + url = "https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg" + + image = Image.open(requests.get(url, stream=True).raw) + image_processor = ViTMAEImageProcessor(size=config.image_size) + inputs = image_processor(images=image, return_tensors="pt") + + # forward pass + torch.manual_seed(2) + outputs = model(**inputs) + logits = outputs.logits + + if "large" in checkpoint_url: + expected_slice = torch.tensor( + [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] + ) + elif "huge" in checkpoint_url: + expected_slice = torch.tensor( + [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] + ) + else: + expected_slice = torch.tensor( + [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] + ) + + # verify logits + assert torch.allclose(logits[0, :3, :3], expected_slice, atol=1e-4) + + print(f"Saving model to {pytorch_dump_folder_path}") + model.save_pretrained(pytorch_dump_folder_path) + + print(f"Saving image processor to {pytorch_dump_folder_path}") + image_processor.save_pretrained(pytorch_dump_folder_path) + + +if __name__ == "__main__": + parser = argparse.ArgumentParser() + # Required parameters + parser.add_argument( + "--checkpoint_url", + default="https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth", + type=str, + help="URL of the checkpoint you'd like to convert.", + ) + parser.add_argument( + "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." + ) + + args = parser.parse_args() + convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path) diff --git a/venv/lib/python3.10/site-packages/transformers/models/vit_mae/modeling_tf_vit_mae.py b/venv/lib/python3.10/site-packages/transformers/models/vit_mae/modeling_tf_vit_mae.py new file mode 100644 index 0000000000000000000000000000000000000000..fff8234e06bc0a080b6bf840da8c8412c82253b5 --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/vit_mae/modeling_tf_vit_mae.py @@ -0,0 +1,1271 @@ +# coding=utf-8 +# Copyright 2022 Facebook AI and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" TF 2.0 ViT MAE (masked autoencoder) model.""" + + +from __future__ import annotations + +import collections.abc +import math +from copy import deepcopy +from dataclasses import dataclass +from typing import Optional, Tuple, Union + +import numpy as np +import tensorflow as tf + +from ...activations_tf import get_tf_activation +from ...file_utils import ( + ModelOutput, + add_start_docstrings, + add_start_docstrings_to_model_forward, + replace_return_docstrings, +) +from ...modeling_tf_outputs import TFBaseModelOutput +from ...modeling_tf_utils import ( + TFModelInputType, + TFPreTrainedModel, + get_initializer, + keras, + keras_serializable, + unpack_inputs, +) +from ...tf_utils import shape_list, stable_softmax +from ...utils import logging +from .configuration_vit_mae import ViTMAEConfig + + +logger = logging.get_logger(__name__) + +_CONFIG_FOR_DOC = "ViTMAEConfig" +_CHECKPOINT_FOR_DOC = "facebook/vit-mae-base" + + +@dataclass +class TFViTMAEModelOutput(ModelOutput): + """ + Class for TFViTMAEModel's outputs, with potential hidden states and attentions. + + Args: + last_hidden_state (`tf.Tensor` of shape `(batch_size, sequence_length, hidden_size)`): + Sequence of hidden-states at the output of the last layer of the model. + mask (`tf.Tensor` of shape `(batch_size, sequence_length)`): + Tensor indicating which patches are masked (1) and which are not (0). + ids_restore (`tf.Tensor` of shape `(batch_size, sequence_length)`): + Tensor containing the original index of the (shuffled) masked patches. + hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape + `(batch_size, sequence_length, hidden_size)`. Hidden-states of the model at the output of each layer plus + the initial embedding outputs. + attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, + sequence_length)`. Attentions weights after the attention softmax, used to compute the weighted average in + the self-attention heads. + """ + + last_hidden_state: tf.Tensor = None + mask: tf.Tensor = None + ids_restore: tf.Tensor = None + hidden_states: Tuple[tf.Tensor] | None = None + attentions: Tuple[tf.Tensor] | None = None + + +@dataclass +class TFViTMAEDecoderOutput(ModelOutput): + """ + Class for TFViTMAEDecoder's outputs, with potential hidden states and attentions. + + Args: + logits (`tf.Tensor` of shape `(batch_size, sequence_length, patch_size ** 2 * num_channels)`): + Pixel reconstruction logits. + hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape + `(batch_size, sequence_length, hidden_size)`. Hidden-states of the model at the output of each layer plus + the initial embedding outputs. + attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, + sequence_length)`. Attentions weights after the attention softmax, used to compute the weighted average in + the self-attention heads. + """ + + logits: tf.Tensor = None + hidden_states: Tuple[tf.Tensor] | None = None + attentions: Tuple[tf.Tensor] | None = None + + +@dataclass +class TFViTMAEForPreTrainingOutput(ModelOutput): + """ + Class for TFViTMAEForPreTraining's outputs, with potential hidden states and attentions. + + Args: + loss (`tf.Tensor` of shape `(1,)`): + Pixel reconstruction loss. + logits (`tf.Tensor` of shape `(batch_size, sequence_length, patch_size ** 2 * num_channels)`): + Pixel reconstruction logits. + mask (`tf.Tensor` of shape `(batch_size, sequence_length)`): + Tensor indicating which patches are masked (1) and which are not (0). + ids_restore (`tf.Tensor` of shape `(batch_size, sequence_length)`): + Tensor containing the original index of the (shuffled) masked patches. + hidden_states (`tuple(tf.Tensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `tf.Tensor` (one for the output of the embeddings + one for the output of each layer) of shape + `(batch_size, sequence_length, hidden_size)`. Hidden-states of the model at the output of each layer plus + the initial embedding outputs. + attentions (`tuple(tf.Tensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `tf.Tensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, + sequence_length)`. Attentions weights after the attention softmax, used to compute the weighted average in + the self-attention heads. + """ + + loss: tf.Tensor | None = None + logits: tf.Tensor = None + mask: tf.Tensor = None + ids_restore: tf.Tensor = None + hidden_states: Tuple[tf.Tensor] | None = None + attentions: Tuple[tf.Tensor] | None = None + + +def get_2d_sincos_pos_embed(embed_dim, grid_size, add_cls_token=False): + """ + Create 2D sin/cos positional embeddings. + + Args: + embed_dim (`int`): + Embedding dimension. + grid_size (`int`): + The grid height and width. + add_cls_token (`bool`, *optional*, defaults to `False`): + Whether or not to add a classification (CLS) token. + + Returns: + (`tf.Tensor` of shape (grid_size*grid_size, embed_dim) or (1+grid_size*grid_size, embed_dim): the position + embeddings (with or without classification token) + """ + grid_h = tf.range(grid_size, dtype=tf.float32) + grid_w = tf.range(grid_size, dtype=tf.float32) + grid = tf.meshgrid(grid_w, grid_h) # here w goes first + grid = tf.stack(grid, axis=0) + + grid = tf.reshape(grid, [2, 1, grid_size, grid_size]) + pos_embed = get_2d_sincos_pos_embed_from_grid(embed_dim, grid) + if add_cls_token: + pos_embed = tf.concat([tf.zeros((1, embed_dim)), pos_embed], axis=0) + return pos_embed + + +def get_2d_sincos_pos_embed_from_grid(embed_dim, grid): + if embed_dim % 2 != 0: + raise ValueError("embed_dim must be even") + + # use half of dimensions to encode grid_h + emb_h = get_1d_sincos_pos_embed_from_grid(embed_dim // 2, grid[0]) # (H*W, D/2) + emb_w = get_1d_sincos_pos_embed_from_grid(embed_dim // 2, grid[1]) # (H*W, D/2) + + emb = tf.concat([emb_h, emb_w], axis=1) # (H*W, D) + return emb + + +def get_1d_sincos_pos_embed_from_grid(embed_dim, pos): + """ + embed_dim: output dimension for each position pos: a list of positions to be encoded: size (M,) out: (M, D) + """ + if embed_dim % 2 != 0: + raise ValueError("embed_dim must be even") + + omega = tf.range(embed_dim // 2, dtype="float32") + omega /= embed_dim / 2.0 + omega = 1.0 / 10000**omega # (D/2,) + + pos = tf.reshape(pos, [-1]) # (M,) + out = tf.einsum("m,d->md", pos, omega) # (M, D/2), outer product + + # half of the positions get sinusoidal pattern and the rest gets + # cosine pattern and then they are concatenated + emb_sin = tf.sin(out) # (M, D/2) + emb_cos = tf.cos(out) # (M, D/2) + + emb = tf.concat([emb_sin, emb_cos], axis=1) # (M, D) + return emb + + +class TFViTMAEEmbeddings(keras.layers.Layer): + """ + Construct the CLS token, position and patch embeddings. + + """ + + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + self.patch_embeddings = TFViTMAEPatchEmbeddings(config, name="patch_embeddings") + self.num_patches = self.patch_embeddings.num_patches + + self.config = config + + def build(self, input_shape=None): + self.cls_token = self.add_weight( + shape=(1, 1, self.config.hidden_size), + initializer=tf.random_normal_initializer(stddev=self.config.initializer_range), + trainable=True, + name="cls_token", + ) + self.position_embeddings = self.add_weight( + shape=(1, self.num_patches + 1, self.config.hidden_size), + initializer="zeros", + trainable=False, # fixed sin-cos embedding + name="position_embeddings", + ) + pos_embed = get_2d_sincos_pos_embed( + self.position_embeddings.shape[-1], + int(self.patch_embeddings.num_patches**0.5), + add_cls_token=True, + )[None, ...] + self.position_embeddings.assign(pos_embed) + + if self.built: + return + self.built = True + if getattr(self, "patch_embeddings", None) is not None: + with tf.name_scope(self.patch_embeddings.name): + self.patch_embeddings.build(None) + + def random_masking(self, sequence: tf.Tensor, noise: tf.Tensor | None = None): + """ + Perform per-sample random masking by per-sample shuffling. Per-sample shuffling is done by argsort random + noise. + + Args: + sequence (`tf.Tensor` of shape `(batch_size, sequence_length, dim)`) + noise (`tf.Tensor` of shape `(batch_size, sequence_length)`, *optional*) which is + mainly used for testing purposes to control randomness and maintain the reproducibility + """ + batch_size, seq_length, dim = shape_list(sequence) + len_keep = int(seq_length * (1 - self.config.mask_ratio)) + + if noise is None: + noise = tf.random.uniform(shape=(batch_size, seq_length), minval=0.0, maxval=1.0) # noise in [0, 1) + + # sort noise for each sample + ids_shuffle = tf.argsort(noise, axis=1) # ascend: small is keep, large is remove + ids_restore = tf.argsort(ids_shuffle, axis=1) + + # keep the first subset + ids_keep = ids_shuffle[:, :len_keep] + sequence_unmasked = tf.gather( + sequence, + axis=1, + batch_dims=1, + indices=ids_keep, + ) + + # generate the binary mask: 0 is keep, 1 is remove + # this hack is needed because TF's EagerTensors don't support + # assignment + mask_keep = tf.zeros((batch_size, len_keep)) + mask_remove = tf.ones((batch_size, seq_length - len_keep)) + mask = tf.concat([mask_keep, mask_remove], axis=-1) + + # unshuffle to get the binary mask + mask = tf.gather(mask, axis=1, batch_dims=1, indices=ids_restore) + + return sequence_unmasked, mask, ids_restore + + def call(self, pixel_values: tf.Tensor, noise: tf.Tensor = None) -> tf.Tensor: + embeddings = self.patch_embeddings(pixel_values) + + # add position embeddings w/o cls token + embeddings = embeddings + self.position_embeddings[:, 1:, :] + + # masking: length -> length * config.mask_ratio + embeddings, mask, ids_restore = self.random_masking(embeddings, noise) + + # append cls token + cls_token = self.cls_token + self.position_embeddings[:, :1, :] + cls_tokens = tf.tile(cls_token, (shape_list(embeddings)[0], 1, 1)) + embeddings = tf.concat([cls_tokens, embeddings], axis=1) + + return embeddings, mask, ids_restore + + +class TFViTMAEPatchEmbeddings(keras.layers.Layer): + """ + This class turns `pixel_values` of shape `(batch_size, num_channels, height, width)` into the initial + `hidden_states` (patch embeddings) of shape `(batch_size, seq_length, hidden_size)` to be consumed by a + Transformer. + """ + + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + image_size, patch_size = config.image_size, config.patch_size + num_channels, hidden_size = config.num_channels, config.hidden_size + image_size = image_size if isinstance(image_size, collections.abc.Iterable) else (image_size, image_size) + patch_size = patch_size if isinstance(patch_size, collections.abc.Iterable) else (patch_size, patch_size) + num_patches = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) + self.image_size = image_size + self.patch_size = patch_size + self.num_patches = num_patches + self.num_channels = num_channels + self.config = config + + self.projection = keras.layers.Conv2D( + filters=hidden_size, + kernel_size=patch_size, + strides=patch_size, + padding="valid", + data_format="channels_last", + kernel_initializer="glorot_uniform", # following torch.nn.Linear + bias_initializer="zeros", + name="projection", + ) + + def call(self, pixel_values: tf.Tensor, training: bool = False) -> tf.Tensor: + batch_size, num_channels, height, width = shape_list(pixel_values) + if tf.executing_eagerly(): + if num_channels != self.num_channels: + raise ValueError( + "Make sure that the channel dimension of the pixel values match with the one set in the" + " configuration." + ) + if height != self.image_size[0] or width != self.image_size[1]: + raise ValueError( + f"Input image size ({height}*{width}) doesn't match model" + f" ({self.image_size[0]}*{self.image_size[1]})." + ) + + # When running on CPU, `keras.layers.Conv2D` doesn't support `NCHW` format. + # So change the input format from `NCHW` to `NHWC`. + # shape = (batch_size, in_height, in_width, in_channels=num_channels) + pixel_values = tf.transpose(pixel_values, perm=(0, 2, 3, 1)) + + projection = self.projection(pixel_values) + + # Change the 2D spatial dimensions to a single temporal dimension. + # shape = (batch_size, num_patches, out_channels=embed_dim) + num_patches = (width // self.patch_size[1]) * (height // self.patch_size[0]) + x = tf.reshape(tensor=projection, shape=(batch_size, num_patches, -1)) + + return x + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "projection", None) is not None: + with tf.name_scope(self.projection.name): + self.projection.build([None, None, None, self.num_channels]) + + +# Copied from transformers.models.vit.modeling_tf_vit.TFViTSelfAttention with ViT->ViTMAE +class TFViTMAESelfAttention(keras.layers.Layer): + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + if config.hidden_size % config.num_attention_heads != 0: + raise ValueError( + f"The hidden size ({config.hidden_size}) is not a multiple of the number " + f"of attention heads ({config.num_attention_heads})" + ) + + self.num_attention_heads = config.num_attention_heads + self.attention_head_size = int(config.hidden_size / config.num_attention_heads) + self.all_head_size = self.num_attention_heads * self.attention_head_size + self.sqrt_att_head_size = math.sqrt(self.attention_head_size) + + self.query = keras.layers.Dense( + units=self.all_head_size, kernel_initializer=get_initializer(config.initializer_range), name="query" + ) + self.key = keras.layers.Dense( + units=self.all_head_size, kernel_initializer=get_initializer(config.initializer_range), name="key" + ) + self.value = keras.layers.Dense( + units=self.all_head_size, kernel_initializer=get_initializer(config.initializer_range), name="value" + ) + self.dropout = keras.layers.Dropout(rate=config.attention_probs_dropout_prob) + self.config = config + + def transpose_for_scores(self, tensor: tf.Tensor, batch_size: int) -> tf.Tensor: + # Reshape from [batch_size, seq_length, all_head_size] to [batch_size, seq_length, num_attention_heads, attention_head_size] + tensor = tf.reshape(tensor=tensor, shape=(batch_size, -1, self.num_attention_heads, self.attention_head_size)) + + # Transpose the tensor from [batch_size, seq_length, num_attention_heads, attention_head_size] to [batch_size, num_attention_heads, seq_length, attention_head_size] + return tf.transpose(tensor, perm=[0, 2, 1, 3]) + + def call( + self, + hidden_states: tf.Tensor, + head_mask: tf.Tensor, + output_attentions: bool, + training: bool = False, + ) -> Tuple[tf.Tensor]: + batch_size = shape_list(hidden_states)[0] + mixed_query_layer = self.query(inputs=hidden_states) + mixed_key_layer = self.key(inputs=hidden_states) + mixed_value_layer = self.value(inputs=hidden_states) + query_layer = self.transpose_for_scores(mixed_query_layer, batch_size) + key_layer = self.transpose_for_scores(mixed_key_layer, batch_size) + value_layer = self.transpose_for_scores(mixed_value_layer, batch_size) + + # Take the dot product between "query" and "key" to get the raw attention scores. + # (batch size, num_heads, seq_len_q, seq_len_k) + attention_scores = tf.matmul(query_layer, key_layer, transpose_b=True) + dk = tf.cast(self.sqrt_att_head_size, dtype=attention_scores.dtype) + attention_scores = tf.divide(attention_scores, dk) + + # Normalize the attention scores to probabilities. + attention_probs = stable_softmax(logits=attention_scores, axis=-1) + + # This is actually dropping out entire tokens to attend to, which might + # seem a bit unusual, but is taken from the original Transformer paper. + attention_probs = self.dropout(inputs=attention_probs, training=training) + + # Mask heads if we want to + if head_mask is not None: + attention_probs = tf.multiply(attention_probs, head_mask) + + attention_output = tf.matmul(attention_probs, value_layer) + attention_output = tf.transpose(attention_output, perm=[0, 2, 1, 3]) + + # (batch_size, seq_len_q, all_head_size) + attention_output = tf.reshape(tensor=attention_output, shape=(batch_size, -1, self.all_head_size)) + outputs = (attention_output, attention_probs) if output_attentions else (attention_output,) + + return outputs + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "query", None) is not None: + with tf.name_scope(self.query.name): + self.query.build([None, None, self.config.hidden_size]) + if getattr(self, "key", None) is not None: + with tf.name_scope(self.key.name): + self.key.build([None, None, self.config.hidden_size]) + if getattr(self, "value", None) is not None: + with tf.name_scope(self.value.name): + self.value.build([None, None, self.config.hidden_size]) + + +# Copied from transformers.models.vit.modeling_tf_vit.TFViTSelfOutput with ViT->ViTMAE +class TFViTMAESelfOutput(keras.layers.Layer): + """ + The residual connection is defined in TFViTMAELayer instead of here (as is the case with other models), due to the + layernorm applied before each block. + """ + + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + self.dense = keras.layers.Dense( + units=config.hidden_size, kernel_initializer=get_initializer(config.initializer_range), name="dense" + ) + self.dropout = keras.layers.Dropout(rate=config.hidden_dropout_prob) + self.config = config + + def call(self, hidden_states: tf.Tensor, input_tensor: tf.Tensor, training: bool = False) -> tf.Tensor: + hidden_states = self.dense(inputs=hidden_states) + hidden_states = self.dropout(inputs=hidden_states, training=training) + + return hidden_states + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "dense", None) is not None: + with tf.name_scope(self.dense.name): + self.dense.build([None, None, self.config.hidden_size]) + + +# Copied from transformers.models.vit.modeling_tf_vit.TFViTAttention with ViT->ViTMAE +class TFViTMAEAttention(keras.layers.Layer): + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + self.self_attention = TFViTMAESelfAttention(config, name="attention") + self.dense_output = TFViTMAESelfOutput(config, name="output") + + def prune_heads(self, heads): + raise NotImplementedError + + def call( + self, + input_tensor: tf.Tensor, + head_mask: tf.Tensor, + output_attentions: bool, + training: bool = False, + ) -> Tuple[tf.Tensor]: + self_outputs = self.self_attention( + hidden_states=input_tensor, head_mask=head_mask, output_attentions=output_attentions, training=training + ) + attention_output = self.dense_output( + hidden_states=self_outputs[0], input_tensor=input_tensor, training=training + ) + outputs = (attention_output,) + self_outputs[1:] # add attentions if we output them + + return outputs + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "self_attention", None) is not None: + with tf.name_scope(self.self_attention.name): + self.self_attention.build(None) + if getattr(self, "dense_output", None) is not None: + with tf.name_scope(self.dense_output.name): + self.dense_output.build(None) + + +# Copied from transformers.models.vit.modeling_tf_vit.TFViTIntermediate with ViT->ViTMAE +class TFViTMAEIntermediate(keras.layers.Layer): + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + self.dense = keras.layers.Dense( + units=config.intermediate_size, kernel_initializer=get_initializer(config.initializer_range), name="dense" + ) + + if isinstance(config.hidden_act, str): + self.intermediate_act_fn = get_tf_activation(config.hidden_act) + else: + self.intermediate_act_fn = config.hidden_act + self.config = config + + def call(self, hidden_states: tf.Tensor) -> tf.Tensor: + hidden_states = self.dense(inputs=hidden_states) + hidden_states = self.intermediate_act_fn(hidden_states) + + return hidden_states + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "dense", None) is not None: + with tf.name_scope(self.dense.name): + self.dense.build([None, None, self.config.hidden_size]) + + +# Copied from transformers.models.vit.modeling_tf_vit.TFViTOutput with ViT->ViTMAE +class TFViTMAEOutput(keras.layers.Layer): + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + self.dense = keras.layers.Dense( + units=config.hidden_size, kernel_initializer=get_initializer(config.initializer_range), name="dense" + ) + self.dropout = keras.layers.Dropout(rate=config.hidden_dropout_prob) + self.config = config + + def call(self, hidden_states: tf.Tensor, input_tensor: tf.Tensor, training: bool = False) -> tf.Tensor: + hidden_states = self.dense(inputs=hidden_states) + hidden_states = self.dropout(inputs=hidden_states, training=training) + hidden_states = hidden_states + input_tensor + + return hidden_states + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "dense", None) is not None: + with tf.name_scope(self.dense.name): + self.dense.build([None, None, self.config.intermediate_size]) + + +# Copied from transformers.models.vit.modeling_tf_vit.TFViTLayer with ViT->ViTMAE +class TFViTMAELayer(keras.layers.Layer): + """This corresponds to the Block class in the timm implementation.""" + + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + self.attention = TFViTMAEAttention(config, name="attention") + self.intermediate = TFViTMAEIntermediate(config, name="intermediate") + self.vit_output = TFViTMAEOutput(config, name="output") + + self.layernorm_before = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="layernorm_before") + self.layernorm_after = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="layernorm_after") + self.config = config + + def call( + self, + hidden_states: tf.Tensor, + head_mask: tf.Tensor, + output_attentions: bool, + training: bool = False, + ) -> Tuple[tf.Tensor]: + attention_outputs = self.attention( + # in ViTMAE, layernorm is applied before self-attention + input_tensor=self.layernorm_before(inputs=hidden_states), + head_mask=head_mask, + output_attentions=output_attentions, + training=training, + ) + attention_output = attention_outputs[0] + + # first residual connection + hidden_states = attention_output + hidden_states + + # in ViTMAE, layernorm is also applied after self-attention + layer_output = self.layernorm_after(inputs=hidden_states) + + intermediate_output = self.intermediate(hidden_states=layer_output) + + # second residual connection is done here + layer_output = self.vit_output( + hidden_states=intermediate_output, input_tensor=hidden_states, training=training + ) + outputs = (layer_output,) + attention_outputs[1:] # add attentions if we output them + + return outputs + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "attention", None) is not None: + with tf.name_scope(self.attention.name): + self.attention.build(None) + if getattr(self, "intermediate", None) is not None: + with tf.name_scope(self.intermediate.name): + self.intermediate.build(None) + if getattr(self, "vit_output", None) is not None: + with tf.name_scope(self.vit_output.name): + self.vit_output.build(None) + if getattr(self, "layernorm_before", None) is not None: + with tf.name_scope(self.layernorm_before.name): + self.layernorm_before.build([None, None, self.config.hidden_size]) + if getattr(self, "layernorm_after", None) is not None: + with tf.name_scope(self.layernorm_after.name): + self.layernorm_after.build([None, None, self.config.hidden_size]) + + +# Copied from transformers.models.vit.modeling_tf_vit.TFViTEncoder with ViT->ViTMAE +class TFViTMAEEncoder(keras.layers.Layer): + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + self.layer = [TFViTMAELayer(config, name=f"layer_._{i}") for i in range(config.num_hidden_layers)] + + def call( + self, + hidden_states: tf.Tensor, + head_mask: tf.Tensor, + output_attentions: bool, + output_hidden_states: bool, + return_dict: bool, + training: bool = False, + ) -> Union[TFBaseModelOutput, Tuple[tf.Tensor]]: + all_hidden_states = () if output_hidden_states else None + all_attentions = () if output_attentions else None + + for i, layer_module in enumerate(self.layer): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + layer_outputs = layer_module( + hidden_states=hidden_states, + head_mask=head_mask[i], + output_attentions=output_attentions, + training=training, + ) + hidden_states = layer_outputs[0] + + if output_attentions: + all_attentions = all_attentions + (layer_outputs[1],) + + # Add last layer + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if not return_dict: + return tuple(v for v in [hidden_states, all_hidden_states, all_attentions] if v is not None) + + return TFBaseModelOutput( + last_hidden_state=hidden_states, hidden_states=all_hidden_states, attentions=all_attentions + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "layer", None) is not None: + for layer in self.layer: + with tf.name_scope(layer.name): + layer.build(None) + + +@keras_serializable +class TFViTMAEMainLayer(keras.layers.Layer): + config_class = ViTMAEConfig + + def __init__(self, config: ViTMAEConfig, **kwargs): + super().__init__(**kwargs) + + self.config = config + + self.embeddings = TFViTMAEEmbeddings(config, name="embeddings") + self.encoder = TFViTMAEEncoder(config, name="encoder") + self.layernorm = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="layernorm") + + def get_input_embeddings(self) -> keras.layers.Layer: + return self.embeddings.patch_embeddings + + def _prune_heads(self, heads_to_prune): + """ + Prunes heads of the model. heads_to_prune: dict of {layer_num: list of heads to prune in this layer} See base + class PreTrainedModel + """ + raise NotImplementedError + + @unpack_inputs + def call( + self, + pixel_values: TFModelInputType | None = None, + noise: tf.Tensor = None, + head_mask: np.ndarray | tf.Tensor | None = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + training: bool = False, + ) -> Union[TFViTMAEModelOutput, Tuple[tf.Tensor]]: + embedding_output, mask, ids_restore = self.embeddings( + pixel_values=pixel_values, training=training, noise=noise + ) + + # Prepare head mask if needed + # 1.0 in head_mask indicate we keep the head + # attention_probs has shape bsz x n_heads x N x N + # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] + # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] + if head_mask is not None: + raise NotImplementedError + else: + head_mask = [None] * self.config.num_hidden_layers + + encoder_outputs = self.encoder( + embedding_output, + head_mask=head_mask, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + + sequence_output = encoder_outputs[0] + sequence_output = self.layernorm(inputs=sequence_output) + + if not return_dict: + return (sequence_output, mask, ids_restore) + encoder_outputs[1:] + + return TFViTMAEModelOutput( + last_hidden_state=sequence_output, + mask=mask, + ids_restore=ids_restore, + hidden_states=encoder_outputs.hidden_states, + attentions=encoder_outputs.attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "embeddings", None) is not None: + with tf.name_scope(self.embeddings.name): + self.embeddings.build(None) + if getattr(self, "encoder", None) is not None: + with tf.name_scope(self.encoder.name): + self.encoder.build(None) + if getattr(self, "layernorm", None) is not None: + with tf.name_scope(self.layernorm.name): + self.layernorm.build([None, None, self.config.hidden_size]) + + +class TFViTMAEPreTrainedModel(TFPreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = ViTMAEConfig + base_model_prefix = "vit" + main_input_name = "pixel_values" + + +VIT_MAE_START_DOCSTRING = r""" + This model inherits from [`TFPreTrainedModel`]. Check the superclass documentation for the generic methods the + library implements for all its model (such as downloading or saving, resizing the input embeddings, pruning heads + etc.) + + This model is also a [keras.Model](https://www.tensorflow.org/api_docs/python/tf/keras/Model) subclass. Use it + as a regular TF 2.0 Keras Model and refer to the TF 2.0 documentation for all matter related to general usage and + behavior. + + + + TensorFlow models and layers in `transformers` accept two formats as input: + + - having all inputs as keyword arguments (like PyTorch models), or + - having all inputs as a list, tuple or dict in the first positional argument. + + The reason the second format is supported is that Keras methods prefer this format when passing inputs to models + and layers. Because of this support, when using methods like `model.fit()` things should "just work" for you - just + pass your inputs and labels in any format that `model.fit()` supports! If, however, you want to use the second + format outside of Keras methods like `fit()` and `predict()`, such as when creating your own layers or models with + the Keras `Functional` API, there are three possibilities you can use to gather all the input Tensors in the first + positional argument: + + - a single Tensor with `pixel_values` only and nothing else: `model(pixel_values)` + - a list of varying length with one or several input Tensors IN THE ORDER given in the docstring: + `model([pixel_values, attention_mask])` or `model([pixel_values, attention_mask, token_type_ids])` + - a dictionary with one or several input Tensors associated to the input names given in the docstring: + `model({"pixel_values": pixel_values, "token_type_ids": token_type_ids})` + + Note that when creating models and layers with + [subclassing](https://keras.io/guides/making_new_layers_and_models_via_subclassing/) then you don't need to worry + about any of this, as you can just pass inputs like you would to any other Python function! + + + + Args: + config ([`ViTMAEConfig`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~TFPreTrainedModel.from_pretrained`] method to load the model weights. +""" + +VIT_MAE_INPUTS_DOCSTRING = r""" + Args: + pixel_values (`np.ndarray`, `tf.Tensor`, `List[tf.Tensor]` ``Dict[str, tf.Tensor]` or `Dict[str, np.ndarray]` and each example must have the shape `(batch_size, num_channels, height, width)`): + Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ViTImageProcessor.__call__`] + for details. + + head_mask (`np.ndarray` or `tf.Tensor` of shape `(num_heads,)` or `(num_layers, num_heads)`, *optional*): + Mask to nullify selected heads of the self-attention modules. Mask values selected in `[0, 1]`: + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. This argument can be used only in eager mode, in graph mode the value in the + config will be used instead. + + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. This argument can be used only in eager mode, in graph mode the value in the config will be + used instead. + + return_dict (`bool`, *optional*): + Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple. This argument can be used + in eager mode, in graph mode the value will always be set to True. + + training (`bool`, *optional*, defaults to `False``): + Whether or not to use the model in training mode (some modules like dropout modules have different + behaviors between training and evaluation). +""" + + +@add_start_docstrings( + "The bare ViTMAE Model transformer outputting raw hidden-states without any specific head on top.", + VIT_MAE_START_DOCSTRING, +) +class TFViTMAEModel(TFViTMAEPreTrainedModel): + def __init__(self, config: ViTMAEConfig, *inputs, **kwargs): + super().__init__(config, *inputs, **kwargs) + + self.vit = TFViTMAEMainLayer(config, name="vit") + + def get_input_embeddings(self): + return self.vit.get_input_embeddings() + + @unpack_inputs + @add_start_docstrings_to_model_forward(VIT_MAE_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=TFViTMAEModelOutput, config_class=_CONFIG_FOR_DOC) + def call( + self, + pixel_values: TFModelInputType | None = None, + noise: tf.Tensor = None, + head_mask: np.ndarray | tf.Tensor | None = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + training: bool = False, + ) -> Union[TFViTMAEModelOutput, Tuple[tf.Tensor]]: + r""" + Returns: + + Examples: + + ```python + >>> from transformers import AutoImageProcessor, TFViTMAEModel + >>> from PIL import Image + >>> import requests + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> image_processor = AutoImageProcessor.from_pretrained("facebook/vit-mae-base") + >>> model = TFViTMAEModel.from_pretrained("facebook/vit-mae-base") + + >>> inputs = image_processor(images=image, return_tensors="tf") + >>> outputs = model(**inputs) + >>> last_hidden_states = outputs.last_hidden_state + ```""" + outputs = self.vit( + pixel_values=pixel_values, + noise=noise, + head_mask=head_mask, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + + return outputs + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "vit", None) is not None: + with tf.name_scope(self.vit.name): + self.vit.build(None) + + +class TFViTMAEDecoder(keras.layers.Layer): + def __init__(self, config, num_patches, **kwargs): + super().__init__(**kwargs) + self.decoder_embed = keras.layers.Dense(config.decoder_hidden_size, name="decoder_embed") + + decoder_config = deepcopy(config) + decoder_config.hidden_size = config.decoder_hidden_size + decoder_config.num_hidden_layers = config.decoder_num_hidden_layers + decoder_config.num_attention_heads = config.decoder_num_attention_heads + decoder_config.intermediate_size = config.decoder_intermediate_size + self.decoder_layers = [ + TFViTMAELayer(decoder_config, name=f"decoder_layers.{j}") for j in range(config.decoder_num_hidden_layers) + ] + + self.decoder_norm = keras.layers.LayerNormalization(epsilon=config.layer_norm_eps, name="decoder_norm") + self.decoder_pred = keras.layers.Dense( + config.patch_size**2 * config.num_channels, + kernel_initializer=get_initializer(config.initializer_range), + name="decoder_pred", + ) # encoder to decoder + self.config = config + self.num_patches = num_patches + + def build(self, input_shape=None): + self.mask_token = self.add_weight( + shape=(1, 1, self.config.decoder_hidden_size), + initializer=tf.random_normal_initializer(stddev=self.config.initializer_range), + trainable=True, + name="mask_token", + ) + self.decoder_pos_embed = self.add_weight( + shape=(1, self.num_patches + 1, self.config.decoder_hidden_size), + initializer="zeros", + trainable=False, + name="decoder_pos_embed", + ) + decoder_pos_embed = get_2d_sincos_pos_embed( + self.decoder_pos_embed.shape[-1], + int(self.num_patches**0.5), + add_cls_token=True, + )[None, ...] + self.decoder_pos_embed.assign(decoder_pos_embed) + + if self.built: + return + self.built = True + if getattr(self, "decoder_embed", None) is not None: + with tf.name_scope(self.decoder_embed.name): + self.decoder_embed.build([None, None, self.config.hidden_size]) + if getattr(self, "decoder_norm", None) is not None: + with tf.name_scope(self.decoder_norm.name): + self.decoder_norm.build([None, None, self.config.decoder_hidden_size]) + if getattr(self, "decoder_pred", None) is not None: + with tf.name_scope(self.decoder_pred.name): + self.decoder_pred.build([None, None, self.config.decoder_hidden_size]) + if getattr(self, "decoder_layers", None) is not None: + for layer in self.decoder_layers: + with tf.name_scope(layer.name): + layer.build(None) + + def call( + self, + hidden_states, + ids_restore, + output_attentions=False, + output_hidden_states=False, + return_dict=True, + ): + # embed tokens + x = self.decoder_embed(hidden_states) + + # append mask tokens to sequence + mask_tokens = tf.tile( + self.mask_token, + (shape_list(x)[0], shape_list(ids_restore)[1] + 1 - shape_list(x)[1], 1), + ) + x_ = tf.concat([x[:, 1:, :], mask_tokens], axis=1) # no cls token + x_ = tf.gather(x_, axis=1, batch_dims=1, indices=ids_restore) # unshuffle + x = tf.concat([x[:, :1, :], x_], axis=1) # append cls token + + # add pos embed + hidden_states = x + self.decoder_pos_embed + + # apply Transformer layers (blocks) + all_hidden_states = () if output_hidden_states else None + all_self_attentions = () if output_attentions else None + for i, layer_module in enumerate(self.decoder_layers): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + layer_outputs = layer_module( + hidden_states, + head_mask=None, + output_attentions=output_attentions, + ) + + hidden_states = layer_outputs[0] + + if output_attentions: + all_self_attentions = all_self_attentions + (layer_outputs[1],) + + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + hidden_states = self.decoder_norm(hidden_states) + + # predictor projection + logits = self.decoder_pred(hidden_states) + + # remove cls token + logits = logits[:, 1:, :] + + if not return_dict: + return tuple(v for v in [logits, all_hidden_states, all_self_attentions] if v is not None) + return TFViTMAEDecoderOutput(logits=logits, hidden_states=all_hidden_states, attentions=all_self_attentions) + + +@add_start_docstrings( + "The ViTMAE Model transformer with the decoder on top for self-supervised pre-training.", + VIT_MAE_START_DOCSTRING, +) +class TFViTMAEForPreTraining(TFViTMAEPreTrainedModel): + def __init__(self, config): + super().__init__(config) + self.config = config + + self.vit = TFViTMAEMainLayer(config, name="vit") + self.decoder = TFViTMAEDecoder( + config, + num_patches=self.vit.embeddings.num_patches, + name="decoder", + ) + + def get_input_embeddings(self): + return self.vit.get_input_embeddings() + + def _prune_heads(self, heads_to_prune): + raise NotImplementedError + + def patchify(self, pixel_values): + """ + Args: + pixel_values (`tf.Tensor` of shape `(batch_size, height, width, num_channels)` or `(batch_size, num_channels, height, width)`): + Pixel values. + + Returns: + `tf.Tensor` of shape `(batch_size, num_patches, patch_size**2 * num_channels)`: + Patchified pixel values. + """ + patch_size, num_channels = self.config.patch_size, self.config.num_channels + # make sure channels are last + if shape_list(pixel_values)[1] == num_channels: + pixel_values = tf.transpose(pixel_values, perm=(0, 2, 3, 1)) + + # sanity checks + tf.debugging.assert_equal( + shape_list(pixel_values)[1], + shape_list(pixel_values)[2], + message="Make sure the pixel values have a squared size", + ) + tf.debugging.assert_equal( + shape_list(pixel_values)[1] % patch_size, + 0, + message="Make sure the pixel values have a size that is divisible by the patch size", + ) + tf.debugging.assert_equal( + shape_list(pixel_values)[3], + num_channels, + message=( + "Make sure the number of channels of the pixel values is equal to the one set in the configuration" + ), + ) + + # patchify + batch_size = shape_list(pixel_values)[0] + num_patches_one_direction = shape_list(pixel_values)[2] // patch_size + patchified_pixel_values = tf.reshape( + pixel_values, + (batch_size, num_patches_one_direction, patch_size, num_patches_one_direction, patch_size, num_channels), + ) + patchified_pixel_values = tf.einsum("nhpwqc->nhwpqc", patchified_pixel_values) + patchified_pixel_values = tf.reshape( + patchified_pixel_values, + (batch_size, num_patches_one_direction * num_patches_one_direction, patch_size**2 * num_channels), + ) + return patchified_pixel_values + + def unpatchify(self, patchified_pixel_values): + """ + Args: + patchified_pixel_values (`tf.Tensor` of shape `(batch_size, num_patches, patch_size**2 * num_channels)`: + Patchified pixel values. + + Returns: + `tf.Tensor` of shape `(batch_size, height, width, num_channels)`: + Pixel values. + """ + patch_size, num_channels = self.config.patch_size, self.config.num_channels + num_patches_one_direction = int(shape_list(patchified_pixel_values)[1] ** 0.5) + # sanity check + tf.debugging.assert_equal( + num_patches_one_direction * num_patches_one_direction, + shape_list(patchified_pixel_values)[1], + message="Make sure that the number of patches can be squared", + ) + + # unpatchify + batch_size = shape_list(patchified_pixel_values)[0] + patchified_pixel_values = tf.reshape( + patchified_pixel_values, + (batch_size, num_patches_one_direction, num_patches_one_direction, patch_size, patch_size, num_channels), + ) + patchified_pixel_values = tf.einsum("nhwpqc->nhpwqc", patchified_pixel_values) + pixel_values = tf.reshape( + patchified_pixel_values, + (batch_size, num_patches_one_direction * patch_size, num_patches_one_direction * patch_size, num_channels), + ) + return pixel_values + + def forward_loss(self, pixel_values, pred, mask): + """ + Args: + pixel_values (`tf.Tensor` of shape `(batch_size, height, width, num_channels)`): + Pixel values. + pred (`tf.Tensor` of shape `(batch_size, num_patches, patch_size**2 * num_channels)`: + Predicted pixel values. + mask (`tf.Tensor` of shape `(batch_size, sequence_length)`): + Tensor indicating which patches are masked (1) and which are not (0). + + Returns: + `tf.Tensor`: Pixel reconstruction loss. + """ + target = self.patchify(pixel_values) + if self.config.norm_pix_loss: + mean = tf.reduce_mean(target, axis=-1, keepdims=True) + var = tf.math.reduce_variance(target, axis=-1, keepdims=True) + target = (target - mean) / (var + 1.0e-6) ** 0.5 + + loss = (pred - target) ** 2 + loss = tf.reduce_mean(loss, axis=-1) # [batch_size, num_patches], mean loss per patch + + loss = tf.reduce_sum(loss * mask) / tf.reduce_sum(mask) # mean loss on removed patches + loss = tf.reshape(loss, (1,)) + return loss + + @unpack_inputs + @add_start_docstrings_to_model_forward(VIT_MAE_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=TFViTMAEForPreTrainingOutput, config_class=_CONFIG_FOR_DOC) + def call( + self, + pixel_values: TFModelInputType | None = None, + noise: tf.Tensor = None, + head_mask: np.ndarray | tf.Tensor | None = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + training: bool = False, + ) -> Union[TFViTMAEForPreTrainingOutput, Tuple[tf.Tensor]]: + r""" + Returns: + + Examples: + + ```python + >>> from transformers import AutoImageProcessor, TFViTMAEForPreTraining + >>> from PIL import Image + >>> import requests + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> image_processor = AutoImageProcessor.from_pretrained("facebook/vit-mae-base") + >>> model = TFViTMAEForPreTraining.from_pretrained("facebook/vit-mae-base") + + >>> inputs = image_processor(images=image, return_tensors="pt") + >>> outputs = model(**inputs) + >>> loss = outputs.loss + >>> mask = outputs.mask + >>> ids_restore = outputs.ids_restore + ```""" + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + outputs = self.vit( + pixel_values=pixel_values, + noise=noise, + head_mask=head_mask, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + training=training, + ) + + latent = outputs.last_hidden_state + ids_restore = outputs.ids_restore + mask = outputs.mask + + decoder_outputs = self.decoder(latent, ids_restore) # [batch_size, num_patches, patch_size**2*3] + logits = decoder_outputs.logits + + loss = self.forward_loss(pixel_values, logits, mask) + + if not return_dict: + output = (logits, mask, ids_restore) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return TFViTMAEForPreTrainingOutput( + loss=loss, + logits=logits, + mask=mask, + ids_restore=ids_restore, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + ) + + def build(self, input_shape=None): + if self.built: + return + self.built = True + if getattr(self, "vit", None) is not None: + with tf.name_scope(self.vit.name): + self.vit.build(None) + if getattr(self, "decoder", None) is not None: + with tf.name_scope(self.decoder.name): + self.decoder.build(None) diff --git a/venv/lib/python3.10/site-packages/transformers/models/vit_mae/modeling_vit_mae.py b/venv/lib/python3.10/site-packages/transformers/models/vit_mae/modeling_vit_mae.py new file mode 100644 index 0000000000000000000000000000000000000000..bfbe59ea903a1a2199afb048b3645606399f053b --- /dev/null +++ b/venv/lib/python3.10/site-packages/transformers/models/vit_mae/modeling_vit_mae.py @@ -0,0 +1,1008 @@ +# coding=utf-8 +# Copyright 2022 Facebook AI and The HuggingFace Inc. team. All rights reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +""" PyTorch ViT MAE (masked autoencoder) model.""" + + +import collections.abc +import math +from copy import deepcopy +from dataclasses import dataclass +from typing import Optional, Set, Tuple, Union + +import numpy as np +import torch +import torch.utils.checkpoint +from torch import nn + +from ...activations import ACT2FN +from ...modeling_outputs import BaseModelOutput +from ...modeling_utils import PreTrainedModel +from ...pytorch_utils import find_pruneable_heads_and_indices, prune_linear_layer +from ...utils import ( + ModelOutput, + add_start_docstrings, + add_start_docstrings_to_model_forward, + logging, + replace_return_docstrings, +) +from .configuration_vit_mae import ViTMAEConfig + + +logger = logging.get_logger(__name__) + +_CONFIG_FOR_DOC = "ViTMAEConfig" +_CHECKPOINT_FOR_DOC = "facebook/vit-mae-base" + + +from ..deprecated._archive_maps import VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST # noqa: F401, E402 + + +@dataclass +class ViTMAEModelOutput(ModelOutput): + """ + Class for ViTMAEModel's outputs, with potential hidden states and attentions. + + Args: + last_hidden_state (`torch.FloatTensor` of shape `(batch_size, sequence_length, hidden_size)`): + Sequence of hidden-states at the output of the last layer of the model. + mask (`torch.FloatTensor` of shape `(batch_size, sequence_length)`): + Tensor indicating which patches are masked (1) and which are not (0). + ids_restore (`torch.LongTensor` of shape `(batch_size, sequence_length)`): + Tensor containing the original index of the (shuffled) masked patches. + hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, sequence_length, hidden_size)`. Hidden-states of the model at the output of each layer + plus the initial embedding outputs. + attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, + sequence_length)`. Attentions weights after the attention softmax, used to compute the weighted average in + the self-attention heads. + """ + + last_hidden_state: torch.FloatTensor = None + mask: torch.LongTensor = None + ids_restore: torch.LongTensor = None + hidden_states: Optional[Tuple[torch.FloatTensor]] = None + attentions: Optional[Tuple[torch.FloatTensor]] = None + + +@dataclass +class ViTMAEDecoderOutput(ModelOutput): + """ + Class for ViTMAEDecoder's outputs, with potential hidden states and attentions. + + Args: + logits (`torch.FloatTensor` of shape `(batch_size, sequence_length, patch_size ** 2 * num_channels)`): + Pixel reconstruction logits. + hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, sequence_length, hidden_size)`. Hidden-states of the model at the output of each layer + plus the initial embedding outputs. + attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, + sequence_length)`. Attentions weights after the attention softmax, used to compute the weighted average in + the self-attention heads. + """ + + logits: torch.FloatTensor = None + hidden_states: Optional[Tuple[torch.FloatTensor]] = None + attentions: Optional[Tuple[torch.FloatTensor]] = None + + +@dataclass +class ViTMAEForPreTrainingOutput(ModelOutput): + """ + Class for ViTMAEForPreTraining's outputs, with potential hidden states and attentions. + + Args: + loss (`torch.FloatTensor` of shape `(1,)`): + Pixel reconstruction loss. + logits (`torch.FloatTensor` of shape `(batch_size, sequence_length, patch_size ** 2 * num_channels)`): + Pixel reconstruction logits. + mask (`torch.FloatTensor` of shape `(batch_size, sequence_length)`): + Tensor indicating which patches are masked (1) and which are not (0). + ids_restore (`torch.LongTensor` of shape `(batch_size, sequence_length)`): + Tensor containing the original index of the (shuffled) masked patches. + hidden_states (`tuple(torch.FloatTensor)`, *optional*, returned when `output_hidden_states=True` is passed or when `config.output_hidden_states=True`): + Tuple of `torch.FloatTensor` (one for the output of the embeddings + one for the output of each layer) of + shape `(batch_size, sequence_length, hidden_size)`. Hidden-states of the model at the output of each layer + plus the initial embedding outputs. + attentions (`tuple(torch.FloatTensor)`, *optional*, returned when `output_attentions=True` is passed or when `config.output_attentions=True`): + Tuple of `torch.FloatTensor` (one for each layer) of shape `(batch_size, num_heads, sequence_length, + sequence_length)`. Attentions weights after the attention softmax, used to compute the weighted average in + the self-attention heads. + """ + + loss: Optional[torch.FloatTensor] = None + logits: torch.FloatTensor = None + mask: torch.LongTensor = None + ids_restore: torch.LongTensor = None + hidden_states: Optional[Tuple[torch.FloatTensor]] = None + attentions: Optional[Tuple[torch.FloatTensor]] = None + + +def get_2d_sincos_pos_embed(embed_dim, grid_size, add_cls_token=False): + """ + Create 2D sin/cos positional embeddings. + + Args: + embed_dim (`int`): + Embedding dimension. + grid_size (`int`): + The grid height and width. + add_cls_token (`bool`, *optional*, defaults to `False`): + Whether or not to add a classification (CLS) token. + + Returns: + (`torch.FloatTensor` of shape (grid_size*grid_size, embed_dim) or (1+grid_size*grid_size, embed_dim): the + position embeddings (with or without classification token) + """ + grid_h = np.arange(grid_size, dtype=np.float32) + grid_w = np.arange(grid_size, dtype=np.float32) + grid = np.meshgrid(grid_w, grid_h) # here w goes first + grid = np.stack(grid, axis=0) + + grid = grid.reshape([2, 1, grid_size, grid_size]) + pos_embed = get_2d_sincos_pos_embed_from_grid(embed_dim, grid) + if add_cls_token: + pos_embed = np.concatenate([np.zeros([1, embed_dim]), pos_embed], axis=0) + return pos_embed + + +def get_2d_sincos_pos_embed_from_grid(embed_dim, grid): + if embed_dim % 2 != 0: + raise ValueError("embed_dim must be even") + + # use half of dimensions to encode grid_h + emb_h = get_1d_sincos_pos_embed_from_grid(embed_dim // 2, grid[0]) # (H*W, D/2) + emb_w = get_1d_sincos_pos_embed_from_grid(embed_dim // 2, grid[1]) # (H*W, D/2) + + emb = np.concatenate([emb_h, emb_w], axis=1) # (H*W, D) + return emb + + +def get_1d_sincos_pos_embed_from_grid(embed_dim, pos): + """ + embed_dim: output dimension for each position pos: a list of positions to be encoded: size (M,) out: (M, D) + """ + if embed_dim % 2 != 0: + raise ValueError("embed_dim must be even") + + omega = np.arange(embed_dim // 2, dtype=float) + omega /= embed_dim / 2.0 + omega = 1.0 / 10000**omega # (D/2,) + + pos = pos.reshape(-1) # (M,) + out = np.einsum("m,d->md", pos, omega) # (M, D/2), outer product + + emb_sin = np.sin(out) # (M, D/2) + emb_cos = np.cos(out) # (M, D/2) + + emb = np.concatenate([emb_sin, emb_cos], axis=1) # (M, D) + return emb + + +class ViTMAEEmbeddings(nn.Module): + """ + Construct the CLS token, position and patch embeddings. + + """ + + def __init__(self, config): + super().__init__() + + self.cls_token = nn.Parameter(torch.zeros(1, 1, config.hidden_size)) + self.patch_embeddings = ViTMAEPatchEmbeddings(config) + self.num_patches = self.patch_embeddings.num_patches + # fixed sin-cos embedding + self.position_embeddings = nn.Parameter( + torch.zeros(1, self.num_patches + 1, config.hidden_size), requires_grad=False + ) + self.config = config + self.initialize_weights() + + def initialize_weights(self): + # initialize (and freeze) position embeddings by sin-cos embedding + pos_embed = get_2d_sincos_pos_embed( + self.position_embeddings.shape[-1], int(self.patch_embeddings.num_patches**0.5), add_cls_token=True + ) + self.position_embeddings.data.copy_(torch.from_numpy(pos_embed).float().unsqueeze(0)) + + # initialize patch_embeddings like nn.Linear (instead of nn.Conv2d) + w = self.patch_embeddings.projection.weight.data + torch.nn.init.xavier_uniform_(w.view([w.shape[0], -1])) + + # timm's trunc_normal_(std=.02) is effectively normal_(std=0.02) as cutoff is too big (2.) + torch.nn.init.normal_(self.cls_token, std=self.config.initializer_range) + + def random_masking(self, sequence, noise=None): + """ + Perform per-sample random masking by per-sample shuffling. Per-sample shuffling is done by argsort random + noise. + + Args: + sequence (`torch.LongTensor` of shape `(batch_size, sequence_length, dim)`) + noise (`torch.FloatTensor` of shape `(batch_size, sequence_length)`, *optional*) which is + mainly used for testing purposes to control randomness and maintain the reproducibility + """ + batch_size, seq_length, dim = sequence.shape + len_keep = int(seq_length * (1 - self.config.mask_ratio)) + + if noise is None: + noise = torch.rand(batch_size, seq_length, device=sequence.device) # noise in [0, 1] + + # sort noise for each sample + ids_shuffle = torch.argsort(noise, dim=1) # ascend: small is keep, large is remove + ids_restore = torch.argsort(ids_shuffle, dim=1) + + # keep the first subset + ids_keep = ids_shuffle[:, :len_keep] + sequence_unmasked = torch.gather(sequence, dim=1, index=ids_keep.unsqueeze(-1).repeat(1, 1, dim)) + + # generate the binary mask: 0 is keep, 1 is remove + mask = torch.ones([batch_size, seq_length], device=sequence.device) + mask[:, :len_keep] = 0 + # unshuffle to get the binary mask + mask = torch.gather(mask, dim=1, index=ids_restore) + + return sequence_unmasked, mask, ids_restore + + def forward(self, pixel_values, noise=None): + batch_size, num_channels, height, width = pixel_values.shape + embeddings = self.patch_embeddings(pixel_values) + + # add position embeddings w/o cls token + embeddings = embeddings + self.position_embeddings[:, 1:, :] + + # masking: length -> length * config.mask_ratio + embeddings, mask, ids_restore = self.random_masking(embeddings, noise) + + # append cls token + cls_token = self.cls_token + self.position_embeddings[:, :1, :] + cls_tokens = cls_token.expand(embeddings.shape[0], -1, -1) + embeddings = torch.cat((cls_tokens, embeddings), dim=1) + + return embeddings, mask, ids_restore + + +class ViTMAEPatchEmbeddings(nn.Module): + """ + This class turns `pixel_values` of shape `(batch_size, num_channels, height, width)` into the initial + `hidden_states` (patch embeddings) of shape `(batch_size, seq_length, hidden_size)` to be consumed by a + Transformer. + """ + + def __init__(self, config): + super().__init__() + image_size, patch_size = config.image_size, config.patch_size + num_channels, hidden_size = config.num_channels, config.hidden_size + image_size = image_size if isinstance(image_size, collections.abc.Iterable) else (image_size, image_size) + patch_size = patch_size if isinstance(patch_size, collections.abc.Iterable) else (patch_size, patch_size) + num_patches = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) + self.image_size = image_size + self.patch_size = patch_size + self.num_channels = num_channels + self.num_patches = num_patches + + self.projection = nn.Conv2d(num_channels, hidden_size, kernel_size=patch_size, stride=patch_size) + + def forward(self, pixel_values): + batch_size, num_channels, height, width = pixel_values.shape + if num_channels != self.num_channels: + raise ValueError( + "Make sure that the channel dimension of the pixel values match with the one set in the configuration." + ) + if height != self.image_size[0] or width != self.image_size[1]: + raise ValueError( + f"Input image size ({height}*{width}) doesn't match model ({self.image_size[0]}*{self.image_size[1]})." + ) + x = self.projection(pixel_values).flatten(2).transpose(1, 2) + return x + + +# Copied from transformers.models.vit.modeling_vit.ViTSelfAttention ViT->ViTMAE +class ViTMAESelfAttention(nn.Module): + def __init__(self, config: ViTMAEConfig) -> None: + super().__init__() + if config.hidden_size % config.num_attention_heads != 0 and not hasattr(config, "embedding_size"): + raise ValueError( + f"The hidden size {config.hidden_size,} is not a multiple of the number of attention " + f"heads {config.num_attention_heads}." + ) + + self.num_attention_heads = config.num_attention_heads + self.attention_head_size = int(config.hidden_size / config.num_attention_heads) + self.all_head_size = self.num_attention_heads * self.attention_head_size + + self.query = nn.Linear(config.hidden_size, self.all_head_size, bias=config.qkv_bias) + self.key = nn.Linear(config.hidden_size, self.all_head_size, bias=config.qkv_bias) + self.value = nn.Linear(config.hidden_size, self.all_head_size, bias=config.qkv_bias) + + self.dropout = nn.Dropout(config.attention_probs_dropout_prob) + + def transpose_for_scores(self, x: torch.Tensor) -> torch.Tensor: + new_x_shape = x.size()[:-1] + (self.num_attention_heads, self.attention_head_size) + x = x.view(new_x_shape) + return x.permute(0, 2, 1, 3) + + def forward( + self, hidden_states, head_mask: Optional[torch.Tensor] = None, output_attentions: bool = False + ) -> Union[Tuple[torch.Tensor, torch.Tensor], Tuple[torch.Tensor]]: + mixed_query_layer = self.query(hidden_states) + + key_layer = self.transpose_for_scores(self.key(hidden_states)) + value_layer = self.transpose_for_scores(self.value(hidden_states)) + query_layer = self.transpose_for_scores(mixed_query_layer) + + # Take the dot product between "query" and "key" to get the raw attention scores. + attention_scores = torch.matmul(query_layer, key_layer.transpose(-1, -2)) + + attention_scores = attention_scores / math.sqrt(self.attention_head_size) + + # Normalize the attention scores to probabilities. + attention_probs = nn.functional.softmax(attention_scores, dim=-1) + + # This is actually dropping out entire tokens to attend to, which might + # seem a bit unusual, but is taken from the original Transformer paper. + attention_probs = self.dropout(attention_probs) + + # Mask heads if we want to + if head_mask is not None: + attention_probs = attention_probs * head_mask + + context_layer = torch.matmul(attention_probs, value_layer) + + context_layer = context_layer.permute(0, 2, 1, 3).contiguous() + new_context_layer_shape = context_layer.size()[:-2] + (self.all_head_size,) + context_layer = context_layer.view(new_context_layer_shape) + + outputs = (context_layer, attention_probs) if output_attentions else (context_layer,) + + return outputs + + +# Copied from transformers.models.vit.modeling_vit.ViTSelfOutput with ViT->ViTMAE +class ViTMAESelfOutput(nn.Module): + """ + The residual connection is defined in ViTMAELayer instead of here (as is the case with other models), due to the + layernorm applied before each block. + """ + + def __init__(self, config: ViTMAEConfig) -> None: + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.hidden_size) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.dropout(hidden_states) + + return hidden_states + + +# Copied from transformers.models.vit.modeling_vit.ViTAttention with ViT->ViTMAE +class ViTMAEAttention(nn.Module): + def __init__(self, config: ViTMAEConfig) -> None: + super().__init__() + self.attention = ViTMAESelfAttention(config) + self.output = ViTMAESelfOutput(config) + self.pruned_heads = set() + + def prune_heads(self, heads: Set[int]) -> None: + if len(heads) == 0: + return + heads, index = find_pruneable_heads_and_indices( + heads, self.attention.num_attention_heads, self.attention.attention_head_size, self.pruned_heads + ) + + # Prune linear layers + self.attention.query = prune_linear_layer(self.attention.query, index) + self.attention.key = prune_linear_layer(self.attention.key, index) + self.attention.value = prune_linear_layer(self.attention.value, index) + self.output.dense = prune_linear_layer(self.output.dense, index, dim=1) + + # Update hyper params and store pruned heads + self.attention.num_attention_heads = self.attention.num_attention_heads - len(heads) + self.attention.all_head_size = self.attention.attention_head_size * self.attention.num_attention_heads + self.pruned_heads = self.pruned_heads.union(heads) + + def forward( + self, + hidden_states: torch.Tensor, + head_mask: Optional[torch.Tensor] = None, + output_attentions: bool = False, + ) -> Union[Tuple[torch.Tensor, torch.Tensor], Tuple[torch.Tensor]]: + self_outputs = self.attention(hidden_states, head_mask, output_attentions) + + attention_output = self.output(self_outputs[0], hidden_states) + + outputs = (attention_output,) + self_outputs[1:] # add attentions if we output them + return outputs + + +# Copied from transformers.models.vit.modeling_vit.ViTIntermediate ViT->ViTMAE +class ViTMAEIntermediate(nn.Module): + def __init__(self, config: ViTMAEConfig) -> None: + super().__init__() + self.dense = nn.Linear(config.hidden_size, config.intermediate_size) + if isinstance(config.hidden_act, str): + self.intermediate_act_fn = ACT2FN[config.hidden_act] + else: + self.intermediate_act_fn = config.hidden_act + + def forward(self, hidden_states: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.intermediate_act_fn(hidden_states) + + return hidden_states + + +# Copied from transformers.models.vit.modeling_vit.ViTOutput ViT->ViTMAE +class ViTMAEOutput(nn.Module): + def __init__(self, config: ViTMAEConfig) -> None: + super().__init__() + self.dense = nn.Linear(config.intermediate_size, config.hidden_size) + self.dropout = nn.Dropout(config.hidden_dropout_prob) + + def forward(self, hidden_states: torch.Tensor, input_tensor: torch.Tensor) -> torch.Tensor: + hidden_states = self.dense(hidden_states) + hidden_states = self.dropout(hidden_states) + + hidden_states = hidden_states + input_tensor + + return hidden_states + + +# Copied from transformers.models.vit.modeling_vit.ViTLayer with ViT->ViTMAE +class ViTMAELayer(nn.Module): + """This corresponds to the Block class in the timm implementation.""" + + def __init__(self, config: ViTMAEConfig) -> None: + super().__init__() + self.chunk_size_feed_forward = config.chunk_size_feed_forward + self.seq_len_dim = 1 + self.attention = ViTMAEAttention(config) + self.intermediate = ViTMAEIntermediate(config) + self.output = ViTMAEOutput(config) + self.layernorm_before = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + self.layernorm_after = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + + def forward( + self, + hidden_states: torch.Tensor, + head_mask: Optional[torch.Tensor] = None, + output_attentions: bool = False, + ) -> Union[Tuple[torch.Tensor, torch.Tensor], Tuple[torch.Tensor]]: + self_attention_outputs = self.attention( + self.layernorm_before(hidden_states), # in ViTMAE, layernorm is applied before self-attention + head_mask, + output_attentions=output_attentions, + ) + attention_output = self_attention_outputs[0] + outputs = self_attention_outputs[1:] # add self attentions if we output attention weights + + # first residual connection + hidden_states = attention_output + hidden_states + + # in ViTMAE, layernorm is also applied after self-attention + layer_output = self.layernorm_after(hidden_states) + layer_output = self.intermediate(layer_output) + + # second residual connection is done here + layer_output = self.output(layer_output, hidden_states) + + outputs = (layer_output,) + outputs + + return outputs + + +# Copied from transformers.models.vit.modeling_vit.ViTEncoder with ViT->ViTMAE +class ViTMAEEncoder(nn.Module): + def __init__(self, config: ViTMAEConfig) -> None: + super().__init__() + self.config = config + self.layer = nn.ModuleList([ViTMAELayer(config) for _ in range(config.num_hidden_layers)]) + self.gradient_checkpointing = False + + def forward( + self, + hidden_states: torch.Tensor, + head_mask: Optional[torch.Tensor] = None, + output_attentions: bool = False, + output_hidden_states: bool = False, + return_dict: bool = True, + ) -> Union[tuple, BaseModelOutput]: + all_hidden_states = () if output_hidden_states else None + all_self_attentions = () if output_attentions else None + + for i, layer_module in enumerate(self.layer): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + layer_head_mask = head_mask[i] if head_mask is not None else None + + if self.gradient_checkpointing and self.training: + layer_outputs = self._gradient_checkpointing_func( + layer_module.__call__, + hidden_states, + layer_head_mask, + output_attentions, + ) + else: + layer_outputs = layer_module(hidden_states, layer_head_mask, output_attentions) + + hidden_states = layer_outputs[0] + + if output_attentions: + all_self_attentions = all_self_attentions + (layer_outputs[1],) + + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if not return_dict: + return tuple(v for v in [hidden_states, all_hidden_states, all_self_attentions] if v is not None) + return BaseModelOutput( + last_hidden_state=hidden_states, + hidden_states=all_hidden_states, + attentions=all_self_attentions, + ) + + +class ViTMAEPreTrainedModel(PreTrainedModel): + """ + An abstract class to handle weights initialization and a simple interface for downloading and loading pretrained + models. + """ + + config_class = ViTMAEConfig + base_model_prefix = "vit" + main_input_name = "pixel_values" + supports_gradient_checkpointing = True + + def _init_weights(self, module): + """Initialize the weights""" + if isinstance(module, (nn.Linear, nn.Conv2d)): + # Slightly different from the TF version which uses truncated_normal for initialization + # cf https://github.com/pytorch/pytorch/pull/5617 + module.weight.data.normal_(mean=0.0, std=self.config.initializer_range) + if module.bias is not None: + module.bias.data.zero_() + elif isinstance(module, nn.LayerNorm): + module.bias.data.zero_() + module.weight.data.fill_(1.0) + + +VIT_MAE_START_DOCSTRING = r""" + This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it + as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and + behavior. + + Parameters: + config ([`ViTMAEConfig`]): Model configuration class with all the parameters of the model. + Initializing with a config file does not load the weights associated with the model, only the + configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights. +""" + +VIT_MAE_INPUTS_DOCSTRING = r""" + Args: + pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): + Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See [`ViTImageProcessor.__call__`] + for details. + + head_mask (`torch.FloatTensor` of shape `(num_heads,)` or `(num_layers, num_heads)`, *optional*): + Mask to nullify selected heads of the self-attention modules. Mask values selected in `[0, 1]`: + + - 1 indicates the head is **not masked**, + - 0 indicates the head is **masked**. + + output_attentions (`bool`, *optional*): + Whether or not to return the attentions tensors of all attention layers. See `attentions` under returned + tensors for more detail. + output_hidden_states (`bool`, *optional*): + Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for + more detail. + return_dict (`bool`, *optional*): + Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple. +""" + + +@add_start_docstrings( + "The bare ViTMAE Model transformer outputting raw hidden-states without any specific head on top.", + VIT_MAE_START_DOCSTRING, +) +class ViTMAEModel(ViTMAEPreTrainedModel): + def __init__(self, config): + super().__init__(config) + self.config = config + + self.embeddings = ViTMAEEmbeddings(config) + self.encoder = ViTMAEEncoder(config) + + self.layernorm = nn.LayerNorm(config.hidden_size, eps=config.layer_norm_eps) + + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self): + return self.embeddings.patch_embeddings + + def _prune_heads(self, heads_to_prune): + """ + Prunes heads of the model. heads_to_prune: dict of {layer_num: list of heads to prune in this layer} See base + class PreTrainedModel + """ + for layer, heads in heads_to_prune.items(): + self.encoder.layer[layer].attention.prune_heads(heads) + + @add_start_docstrings_to_model_forward(VIT_MAE_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=ViTMAEModelOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + pixel_values: Optional[torch.FloatTensor] = None, + noise: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, ViTMAEModelOutput]: + r""" + Returns: + + Examples: + + ```python + >>> from transformers import AutoImageProcessor, ViTMAEModel + >>> from PIL import Image + >>> import requests + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> image_processor = AutoImageProcessor.from_pretrained("facebook/vit-mae-base") + >>> model = ViTMAEModel.from_pretrained("facebook/vit-mae-base") + + >>> inputs = image_processor(images=image, return_tensors="pt") + >>> outputs = model(**inputs) + >>> last_hidden_states = outputs.last_hidden_state + ```""" + output_attentions = output_attentions if output_attentions is not None else self.config.output_attentions + output_hidden_states = ( + output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states + ) + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + if pixel_values is None: + raise ValueError("You have to specify pixel_values") + + # Prepare head mask if needed + # 1.0 in head_mask indicate we keep the head + # attention_probs has shape bsz x n_heads x N x N + # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] + # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] + head_mask = self.get_head_mask(head_mask, self.config.num_hidden_layers) + + embedding_output, mask, ids_restore = self.embeddings(pixel_values, noise=noise) + + encoder_outputs = self.encoder( + embedding_output, + head_mask=head_mask, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + sequence_output = encoder_outputs[0] + sequence_output = self.layernorm(sequence_output) + + if not return_dict: + return (sequence_output, mask, ids_restore) + encoder_outputs[1:] + + return ViTMAEModelOutput( + last_hidden_state=sequence_output, + mask=mask, + ids_restore=ids_restore, + hidden_states=encoder_outputs.hidden_states, + attentions=encoder_outputs.attentions, + ) + + +class ViTMAEDecoder(nn.Module): + def __init__(self, config, num_patches): + super().__init__() + self.decoder_embed = nn.Linear(config.hidden_size, config.decoder_hidden_size, bias=True) + self.mask_token = nn.Parameter(torch.zeros(1, 1, config.decoder_hidden_size)) + self.decoder_pos_embed = nn.Parameter( + torch.zeros(1, num_patches + 1, config.decoder_hidden_size), requires_grad=False + ) # fixed sin-cos embedding + + decoder_config = deepcopy(config) + decoder_config.hidden_size = config.decoder_hidden_size + decoder_config.num_hidden_layers = config.decoder_num_hidden_layers + decoder_config.num_attention_heads = config.decoder_num_attention_heads + decoder_config.intermediate_size = config.decoder_intermediate_size + self.decoder_layers = nn.ModuleList( + [ViTMAELayer(decoder_config) for _ in range(config.decoder_num_hidden_layers)] + ) + + self.decoder_norm = nn.LayerNorm(config.decoder_hidden_size, eps=config.layer_norm_eps) + self.decoder_pred = nn.Linear( + config.decoder_hidden_size, config.patch_size**2 * config.num_channels, bias=True + ) # encoder to decoder + self.gradient_checkpointing = False + self.config = config + self.initialize_weights(num_patches) + + def initialize_weights(self, num_patches): + # initialize (and freeze) position embeddings by sin-cos embedding + decoder_pos_embed = get_2d_sincos_pos_embed( + self.decoder_pos_embed.shape[-1], int(num_patches**0.5), add_cls_token=True + ) + self.decoder_pos_embed.data.copy_(torch.from_numpy(decoder_pos_embed).float().unsqueeze(0)) + + # timm's trunc_normal_(std=.02) is effectively normal_(std=0.02) as cutoff is too big (2.) + torch.nn.init.normal_(self.mask_token, std=self.config.initializer_range) + + def forward( + self, + hidden_states, + ids_restore, + output_attentions=False, + output_hidden_states=False, + return_dict=True, + ): + # embed tokens + x = self.decoder_embed(hidden_states) + + # append mask tokens to sequence + mask_tokens = self.mask_token.repeat(x.shape[0], ids_restore.shape[1] + 1 - x.shape[1], 1) + x_ = torch.cat([x[:, 1:, :], mask_tokens], dim=1) # no cls token + x_ = torch.gather(x_, dim=1, index=ids_restore.unsqueeze(-1).repeat(1, 1, x.shape[2])) # unshuffle + x = torch.cat([x[:, :1, :], x_], dim=1) # append cls token + + # add pos embed + hidden_states = x + self.decoder_pos_embed + + # apply Transformer layers (blocks) + all_hidden_states = () if output_hidden_states else None + all_self_attentions = () if output_attentions else None + for i, layer_module in enumerate(self.decoder_layers): + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + if self.gradient_checkpointing and self.training: + layer_outputs = self._gradient_checkpointing_func( + layer_module.__call__, + hidden_states, + None, + output_attentions, + ) + else: + layer_outputs = layer_module(hidden_states, head_mask=None, output_attentions=output_attentions) + + hidden_states = layer_outputs[0] + + if output_attentions: + all_self_attentions = all_self_attentions + (layer_outputs[1],) + + if output_hidden_states: + all_hidden_states = all_hidden_states + (hidden_states,) + + hidden_states = self.decoder_norm(hidden_states) + + # predictor projection + logits = self.decoder_pred(hidden_states) + + # remove cls token + logits = logits[:, 1:, :] + + if not return_dict: + return tuple(v for v in [logits, all_hidden_states, all_self_attentions] if v is not None) + return ViTMAEDecoderOutput( + logits=logits, + hidden_states=all_hidden_states, + attentions=all_self_attentions, + ) + + +@add_start_docstrings( + """The ViTMAE Model transformer with the decoder on top for self-supervised pre-training. + + + + Note that we provide a script to pre-train this model on custom data in our [examples + directory](https://github.com/huggingface/transformers/tree/main/examples/pytorch/image-pretraining). + + + + """, + VIT_MAE_START_DOCSTRING, +) +class ViTMAEForPreTraining(ViTMAEPreTrainedModel): + def __init__(self, config): + super().__init__(config) + self.config = config + + self.vit = ViTMAEModel(config) + self.decoder = ViTMAEDecoder(config, num_patches=self.vit.embeddings.num_patches) + + # Initialize weights and apply final processing + self.post_init() + + def get_input_embeddings(self): + return self.vit.embeddings.patch_embeddings + + def _prune_heads(self, heads_to_prune): + """ + Prunes heads of the model. heads_to_prune: dict of {layer_num: list of heads to prune in this layer} See base + class PreTrainedModel + """ + for layer, heads in heads_to_prune.items(): + self.encoder.layer[layer].attention.prune_heads(heads) + + def patchify(self, pixel_values): + """ + Args: + pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): + Pixel values. + + Returns: + `torch.FloatTensor` of shape `(batch_size, num_patches, patch_size**2 * num_channels)`: + Patchified pixel values. + """ + patch_size, num_channels = self.config.patch_size, self.config.num_channels + # sanity checks + if (pixel_values.shape[2] != pixel_values.shape[3]) or (pixel_values.shape[2] % patch_size != 0): + raise ValueError("Make sure the pixel values have a squared size that is divisible by the patch size") + if pixel_values.shape[1] != num_channels: + raise ValueError( + "Make sure the number of channels of the pixel values is equal to the one set in the configuration" + ) + + # patchify + batch_size = pixel_values.shape[0] + num_patches_one_direction = pixel_values.shape[2] // patch_size + patchified_pixel_values = pixel_values.reshape( + batch_size, num_channels, num_patches_one_direction, patch_size, num_patches_one_direction, patch_size + ) + patchified_pixel_values = torch.einsum("nchpwq->nhwpqc", patchified_pixel_values) + patchified_pixel_values = patchified_pixel_values.reshape( + batch_size, num_patches_one_direction * num_patches_one_direction, patch_size**2 * num_channels + ) + return patchified_pixel_values + + def unpatchify(self, patchified_pixel_values): + """ + Args: + patchified_pixel_values (`torch.FloatTensor` of shape `(batch_size, num_patches, patch_size**2 * num_channels)`: + Patchified pixel values. + + Returns: + `torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`: + Pixel values. + """ + patch_size, num_channels = self.config.patch_size, self.config.num_channels + num_patches_one_direction = int(patchified_pixel_values.shape[1] ** 0.5) + # sanity check + if num_patches_one_direction**2 != patchified_pixel_values.shape[1]: + raise ValueError("Make sure that the number of patches can be squared") + + # unpatchify + batch_size = patchified_pixel_values.shape[0] + patchified_pixel_values = patchified_pixel_values.reshape( + batch_size, + num_patches_one_direction, + num_patches_one_direction, + patch_size, + patch_size, + num_channels, + ) + patchified_pixel_values = torch.einsum("nhwpqc->nchpwq", patchified_pixel_values) + pixel_values = patchified_pixel_values.reshape( + batch_size, + num_channels, + num_patches_one_direction * patch_size, + num_patches_one_direction * patch_size, + ) + return pixel_values + + def forward_loss(self, pixel_values, pred, mask): + """ + Args: + pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`): + Pixel values. + pred (`torch.FloatTensor` of shape `(batch_size, num_patches, patch_size**2 * num_channels)`: + Predicted pixel values. + mask (`torch.FloatTensor` of shape `(batch_size, sequence_length)`): + Tensor indicating which patches are masked (1) and which are not (0). + + Returns: + `torch.FloatTensor`: Pixel reconstruction loss. + """ + target = self.patchify(pixel_values) + if self.config.norm_pix_loss: + mean = target.mean(dim=-1, keepdim=True) + var = target.var(dim=-1, keepdim=True) + target = (target - mean) / (var + 1.0e-6) ** 0.5 + + loss = (pred - target) ** 2 + loss = loss.mean(dim=-1) # [N, L], mean loss per patch + + loss = (loss * mask).sum() / mask.sum() # mean loss on removed patches + return loss + + @add_start_docstrings_to_model_forward(VIT_MAE_INPUTS_DOCSTRING) + @replace_return_docstrings(output_type=ViTMAEForPreTrainingOutput, config_class=_CONFIG_FOR_DOC) + def forward( + self, + pixel_values: Optional[torch.FloatTensor] = None, + noise: Optional[torch.FloatTensor] = None, + head_mask: Optional[torch.FloatTensor] = None, + output_attentions: Optional[bool] = None, + output_hidden_states: Optional[bool] = None, + return_dict: Optional[bool] = None, + ) -> Union[Tuple, ViTMAEForPreTrainingOutput]: + r""" + Returns: + + Examples: + + ```python + >>> from transformers import AutoImageProcessor, ViTMAEForPreTraining + >>> from PIL import Image + >>> import requests + + >>> url = "http://images.cocodataset.org/val2017/000000039769.jpg" + >>> image = Image.open(requests.get(url, stream=True).raw) + + >>> image_processor = AutoImageProcessor.from_pretrained("facebook/vit-mae-base") + >>> model = ViTMAEForPreTraining.from_pretrained("facebook/vit-mae-base") + + >>> inputs = image_processor(images=image, return_tensors="pt") + >>> outputs = model(**inputs) + >>> loss = outputs.loss + >>> mask = outputs.mask + >>> ids_restore = outputs.ids_restore + ```""" + return_dict = return_dict if return_dict is not None else self.config.use_return_dict + + outputs = self.vit( + pixel_values, + noise=noise, + head_mask=head_mask, + output_attentions=output_attentions, + output_hidden_states=output_hidden_states, + return_dict=return_dict, + ) + + latent = outputs.last_hidden_state + ids_restore = outputs.ids_restore + mask = outputs.mask + + decoder_outputs = self.decoder(latent, ids_restore) + logits = decoder_outputs.logits # shape (batch_size, num_patches, patch_size*patch_size*num_channels) + + loss = self.forward_loss(pixel_values, logits, mask) + + if not return_dict: + output = (logits, mask, ids_restore) + outputs[2:] + return ((loss,) + output) if loss is not None else output + + return ViTMAEForPreTrainingOutput( + loss=loss, + logits=logits, + mask=mask, + ids_restore=ids_restore, + hidden_states=outputs.hidden_states, + attentions=outputs.attentions, + )