diff --git a/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5c1ac4e49694bcc5d68ae5be2d5f70da2044a198 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715686900 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..e490edecb74ece2a51d83347cc4506ea3caa35a7 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log @@ -0,0 +1,28 @@ + +2024-05-14:11:41:40,928 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:11:41:46,726 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:11:41:46,729 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:11:41:46,729 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step100'} +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/core/register.py:145: UserWarning: "hpu:X" notation is not supported by Gaudi PyTorch intergration bridge. Please change to "hpu" without index (Triggered internally at /npu-stack/pytorch-integration/pytorch_helpers/lazy_to_backend.cpp:53.) + return func(*args, **kwargs) +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +[2024-05-14 11:41:58,467] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +2024-05-14:11:41:58,855 WARNING [task.py:763] [Task: indiccopa-hi] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-14:11:41:58,855 WARNING [task.py:775] [Task: indiccopa-hi] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for ai4bharat/IndicCOPA contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/ai4bharat/IndicCOPA +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-05-14:11:42:00,273 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:11:42:00,273 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:11:42:00,292 INFO [task.py:395] Building contexts for indiccopa-hi on rank 2... +100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 56/56 [00:00<00:00, 102255.56it/s] +Passed argument batch_size = auto:1. Detecting largest batch size +Determined largest batch size: 64 +2024-05-14:11:42:01,857 INFO [evaluator.py:379] Running loglikelihood requests +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..7be7472b69b47081a3c31e787a92de060a9cd0a3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T11:41:40.738216", + "startedAt": "2024-05-14T11:41:40.263939", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=global_step100" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3387.3559407894736, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3226.855, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3226.846, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3203.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3226.836, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3286.754, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.003, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3034.448, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3219.993, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3215.612, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3289.224, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3295.916, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3316.248, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 77.68973922729492 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..4841bc4504e219db83b702d09e68cfaa8fa95063 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 28}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..7392164b3f3165524cc7e53e82310eb7e0716bca --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/logs/debug-internal.log @@ -0,0 +1,194 @@ +2024-05-14 11:41:40,280 INFO StreamThr :72092 [internal.py:wandb_internal():85] W&B internal server running at pid: 72092, started at: 2024-05-14 11:41:40.279266 +2024-05-14 11:41:40,281 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: status +2024-05-14 11:41:40,282 INFO WriterThread:72092 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/run-cn9l5r80.wandb +2024-05-14 11:41:40,284 DEBUG SenderThread:72092 [sender.py:send():378] send: header +2024-05-14 11:41:40,294 DEBUG SenderThread:72092 [sender.py:send():378] send: run +2024-05-14 11:41:40,562 INFO SenderThread:72092 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files +2024-05-14 11:41:40,562 INFO SenderThread:72092 [sender.py:_start_run_threads():1123] run started: cn9l5r80 with start time 1715686900.278444 +2024-05-14 11:41:40,568 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 11:41:40,568 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: check_version +2024-05-14 11:41:40,648 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 11:41:40,649 DEBUG HandlerThread:72092 [system_info.py:__init__():26] System info init +2024-05-14 11:41:40,649 DEBUG HandlerThread:72092 [system_info.py:__init__():41] System info init done +2024-05-14 11:41:40,650 INFO HandlerThread:72092 [system_monitor.py:start():194] Starting system monitor +2024-05-14 11:41:40,650 INFO SystemMonitor:72092 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 11:41:40,650 INFO HandlerThread:72092 [system_monitor.py:probe():214] Collecting system info +2024-05-14 11:41:40,650 INFO SystemMonitor:72092 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 11:41:40,651 INFO SystemMonitor:72092 [interfaces.py:start():188] Started disk monitoring +2024-05-14 11:41:40,652 INFO SystemMonitor:72092 [interfaces.py:start():188] Started memory monitoring +2024-05-14 11:41:40,652 INFO SystemMonitor:72092 [interfaces.py:start():188] Started network monitoring +2024-05-14 11:41:40,738 DEBUG HandlerThread:72092 [system_info.py:probe():150] Probing system +2024-05-14 11:41:40,791 DEBUG HandlerThread:72092 [system_info.py:_probe_git():135] Probing git +2024-05-14 11:41:40,816 ERROR HandlerThread:72092 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 11:41:40,817 DEBUG HandlerThread:72092 [system_info.py:_probe_git():143] Probing git done +2024-05-14 11:41:40,817 DEBUG HandlerThread:72092 [system_info.py:probe():198] Probing system done +2024-05-14 11:41:40,817 DEBUG HandlerThread:72092 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T11:41:40.738216', 'startedAt': '2024-05-14T11:41:40.263939', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=global_step100'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3387.3559407894736, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3226.855, 'min': 800.0, 'max': 3400.0}, {'current': 3226.846, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3203.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3226.836, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3286.754, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.003, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3034.448, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3219.993, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3215.612, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3289.224, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3295.916, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3316.248, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 77.68973922729492}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 11:41:40,817 INFO HandlerThread:72092 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 11:41:40,817 INFO HandlerThread:72092 [system_monitor.py:probe():227] Publishing system info +2024-05-14 11:41:40,818 INFO HandlerThread:72092 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 11:41:40,822 DEBUG SenderThread:72092 [sender.py:send():378] send: files +2024-05-14 11:41:40,822 INFO SenderThread:72092 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 11:41:40,924 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 11:41:40,924 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 11:41:40,924 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: python_packages +2024-05-14 11:41:40,925 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: stop_status +2024-05-14 11:41:41,144 DEBUG SenderThread:72092 [sender.py:send():378] send: telemetry +2024-05-14 11:41:41,334 INFO wandb-upload_0:72092 [upload_job.py:push():130] Uploaded file /tmp/tmpvm9dypaiwandb/tibjvxil-wandb-metadata.json +2024-05-14 11:41:41,564 INFO Thread-12 :72092 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-metadata.json +2024-05-14 11:41:41,564 INFO Thread-12 :72092 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:41:41,564 INFO Thread-12 :72092 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/requirements.txt +2024-05-14 11:41:43,563 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:41:46,146 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:41:49,570 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:41:51,731 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:41:55,575 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:41:55,925 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 11:41:55,925 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: stop_status +2024-05-14 11:41:57,011 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:41:59,578 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:42:00,579 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:42:01,579 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:42:02,362 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:42:02,580 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:42:03,582 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:42:07,363 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:42:09,383 DEBUG SenderThread:72092 [sender.py:send():378] send: exit +2024-05-14 11:42:09,383 INFO SenderThread:72092 [sender.py:send_exit():585] handling exit code: 0 +2024-05-14 11:42:09,384 INFO SenderThread:72092 [sender.py:send_exit():587] handling runtime: 28 +2024-05-14 11:42:09,384 INFO SenderThread:72092 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 11:42:09,385 INFO SenderThread:72092 [sender.py:send_exit():593] send defer +2024-05-14 11:42:09,385 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:09,385 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 11:42:09,385 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:09,385 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 11:42:09,385 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 1 +2024-05-14 11:42:09,385 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:09,385 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 11:42:09,385 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:09,385 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 11:42:09,385 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 2 +2024-05-14 11:42:09,385 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:09,385 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 11:42:09,385 INFO HandlerThread:72092 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 11:42:09,386 DEBUG SystemMonitor:72092 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 11:42:09,386 INFO HandlerThread:72092 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 11:42:09,386 DEBUG SystemMonitor:72092 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 11:42:09,386 INFO HandlerThread:72092 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 11:42:09,386 DEBUG SystemMonitor:72092 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 11:42:09,386 INFO HandlerThread:72092 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 11:42:09,388 INFO HandlerThread:72092 [interfaces.py:finish():200] Joined network monitor +2024-05-14 11:42:09,388 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:09,388 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 11:42:09,388 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 3 +2024-05-14 11:42:09,388 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:09,388 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 11:42:09,389 DEBUG SenderThread:72092 [sender.py:send():378] send: stats +2024-05-14 11:42:09,389 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:09,389 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 11:42:09,389 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 4 +2024-05-14 11:42:09,389 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:09,389 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 11:42:09,389 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:09,389 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 11:42:09,389 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 5 +2024-05-14 11:42:09,390 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:09,390 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 11:42:09,390 DEBUG SenderThread:72092 [sender.py:send():378] send: summary +2024-05-14 11:42:09,390 INFO SenderThread:72092 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 11:42:09,390 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:09,391 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 11:42:09,391 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 6 +2024-05-14 11:42:09,391 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:09,391 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 11:42:09,391 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:09,391 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 11:42:09,393 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 11:42:09,496 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 7 +2024-05-14 11:42:09,496 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:09,496 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 11:42:09,497 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:09,497 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 11:42:09,590 INFO Thread-12 :72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/config.yaml +2024-05-14 11:42:09,590 INFO Thread-12 :72092 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-summary.json +2024-05-14 11:42:10,384 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 11:42:12,483 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 8 +2024-05-14 11:42:12,483 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 11:42:12,483 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:12,483 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 11:42:12,484 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:12,484 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 11:42:12,484 INFO SenderThread:72092 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 11:42:12,484 INFO SenderThread:72092 [job_builder.py:_get_source_type():576] no source found +2024-05-14 11:42:12,484 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 9 +2024-05-14 11:42:12,484 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:12,484 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 11:42:12,484 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:12,484 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 11:42:12,484 INFO SenderThread:72092 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 11:42:12,592 INFO SenderThread:72092 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:42:12,592 INFO SenderThread:72092 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files +2024-05-14 11:42:12,593 INFO SenderThread:72092 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-metadata.json wandb-metadata.json +2024-05-14 11:42:12,593 INFO SenderThread:72092 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/config.yaml config.yaml +2024-05-14 11:42:12,593 INFO SenderThread:72092 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/requirements.txt requirements.txt +2024-05-14 11:42:12,593 INFO SenderThread:72092 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log output.log +2024-05-14 11:42:12,595 INFO SenderThread:72092 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-summary.json wandb-summary.json +2024-05-14 11:42:12,595 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 10 +2024-05-14 11:42:12,596 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:12,597 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 11:42:12,599 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:12,599 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 11:42:12,599 INFO SenderThread:72092 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 11:42:12,829 INFO wandb-upload_1:72092 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/requirements.txt +2024-05-14 11:42:13,023 INFO wandb-upload_0:72092 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/config.yaml +2024-05-14 11:42:13,089 INFO wandb-upload_3:72092 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/wandb-summary.json +2024-05-14 11:42:13,099 INFO wandb-upload_2:72092 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/files/output.log +2024-05-14 11:42:13,299 INFO Thread-11 (_thread_body):72092 [sender.py:transition_state():613] send defer: 11 +2024-05-14 11:42:13,300 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:13,300 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 11:42:13,300 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:13,300 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 11:42:13,300 INFO SenderThread:72092 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 11:42:13,301 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 12 +2024-05-14 11:42:13,301 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:13,301 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 11:42:13,301 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:13,301 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 11:42:13,301 INFO SenderThread:72092 [file_stream.py:finish():601] file stream finish called +2024-05-14 11:42:13,384 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 11:42:13,542 INFO SenderThread:72092 [file_stream.py:finish():605] file stream finish is done +2024-05-14 11:42:13,542 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 13 +2024-05-14 11:42:13,542 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 11:42:13,542 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:13,542 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 11:42:13,543 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:13,543 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 11:42:13,543 INFO SenderThread:72092 [sender.py:transition_state():613] send defer: 14 +2024-05-14 11:42:13,543 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: defer +2024-05-14 11:42:13,543 INFO HandlerThread:72092 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 11:42:13,543 DEBUG SenderThread:72092 [sender.py:send():378] send: final +2024-05-14 11:42:13,543 DEBUG SenderThread:72092 [sender.py:send():378] send: footer +2024-05-14 11:42:13,543 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: defer +2024-05-14 11:42:13,543 INFO SenderThread:72092 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 11:42:13,543 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 11:42:13,543 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 11:42:13,543 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 11:42:13,544 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 11:42:13,544 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 11:42:13,544 DEBUG SenderThread:72092 [sender.py:send_request():405] send_request: server_info +2024-05-14 11:42:13,545 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 11:42:13,545 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 11:42:13,546 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 11:42:13,607 INFO MainThread:72092 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 11:42:13,607 INFO MainThread:72092 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 11:42:13,607 INFO MainThread:72092 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 11:42:13,607 DEBUG HandlerThread:72092 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 11:42:13,608 INFO HandlerThread:72092 [handler.py:finish():882] shutting down handler +2024-05-14 11:42:14,544 INFO WriterThread:72092 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/run-cn9l5r80.wandb +2024-05-14 11:42:14,607 INFO SenderThread:72092 [sender.py:finish():1545] shutting down sender +2024-05-14 11:42:14,607 INFO SenderThread:72092 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 11:42:14,607 INFO SenderThread:72092 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c57fdff5f110734a37e11646253bd280dab15c70 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_setup.py:_flush():76] Configure stats pid to 70754 +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 11:41:40,275 WARNING MainThread:70754 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/logs/debug.log +2024-05-14 11:41:40,275 INFO MainThread:70754 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/logs/debug-internal.log +2024-05-14 11:41:40,276 INFO MainThread:70754 [wandb_init.py:init():560] calling init triggers +2024-05-14 11:41:40,276 INFO MainThread:70754 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 11:41:40,276 INFO MainThread:70754 [wandb_init.py:init():610] starting backend +2024-05-14 11:41:40,276 INFO MainThread:70754 [wandb_init.py:init():614] setting up manager +2024-05-14 11:41:40,277 INFO MainThread:70754 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 11:41:40,278 INFO MainThread:70754 [wandb_init.py:init():622] backend started and connected +2024-05-14 11:41:40,282 INFO MainThread:70754 [wandb_init.py:init():711] updated telemetry +2024-05-14 11:41:40,293 INFO MainThread:70754 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 11:41:40,568 INFO MainThread:70754 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 11:41:40,644 INFO MainThread:70754 [wandb_run.py:_on_init():2405] got version response +2024-05-14 11:41:40,644 INFO MainThread:70754 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 11:41:40,925 INFO MainThread:70754 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 11:41:40,925 INFO MainThread:70754 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 11:41:40,925 INFO MainThread:70754 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 11:41:40,925 INFO MainThread:70754 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 11:41:40,926 INFO MainThread:70754 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 11:42:14,608 WARNING MsgRouterThr:70754 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/run-cn9l5r80.wandb b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/run-cn9l5r80.wandb new file mode 100644 index 0000000000000000000000000000000000000000..1105a302342fac03867df9183c72e9a2c1b11d6e Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_114140-cn9l5r80/run-cn9l5r80.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/files/output.log b/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4f2e6d58e2352699a4e863a3e0cb6246cef5526e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/files/output.log @@ -0,0 +1,28 @@ + +2024-05-14:11:44:44,817 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:11:44:49,811 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:11:44:49,814 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:11:44:49,814 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step100'} +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/core/register.py:145: UserWarning: "hpu:X" notation is not supported by Gaudi PyTorch intergration bridge. Please change to "hpu" without index (Triggered internally at /npu-stack/pytorch-integration/pytorch_helpers/lazy_to_backend.cpp:53.) + return func(*args, **kwargs) +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +2024-05-14:11:44:58,230 WARNING [task.py:763] [Task: indiccopa-hi] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-14:11:44:58,230 WARNING [task.py:775] [Task: indiccopa-hi] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +[2024-05-14 11:44:57,821] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for ai4bharat/IndicCOPA contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/ai4bharat/IndicCOPA +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-05-14:11:45:03,646 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:11:45:03,646 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:11:45:03,668 INFO [task.py:395] Building contexts for indiccopa-hi on rank 7... +100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 56/56 [00:00<00:00, 108540.21it/s] +2024-05-14:11:45:05,430 INFO [evaluator.py:379] Running loglikelihood requests +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Passed argument batch_size = auto:1. Detecting largest batch size +Determined largest batch size: 64 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..f97bfbe43401ee851b2b689eb650f147f8bb41d3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 29}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/run-ht38n8wu.wandb b/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/run-ht38n8wu.wandb new file mode 100644 index 0000000000000000000000000000000000000000..872e6cf435f9928b781ab77349b650c6d80c2296 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_114444-ht38n8wu/run-ht38n8wu.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c9f6c8298d7f6388c49b502891050f1a4aef6d7c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715705031 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..8420f6737726cbcb95ffbbea76e92ce606e27166 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log @@ -0,0 +1,28 @@ + +2024-05-14:16:43:51,783 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:16:43:56,475 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:16:43:56,477 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:16:43:56,477 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step100'} +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/core/register.py:145: UserWarning: "hpu:X" notation is not supported by Gaudi PyTorch intergration bridge. Please change to "hpu" without index (Triggered internally at /npu-stack/pytorch-integration/pytorch_helpers/lazy_to_backend.cpp:53.) + return func(*args, **kwargs) +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/hpu/__init__.py:158: UserWarning: torch.hpu.setDeterministic is deprecated and will be removed in next release. Please use torch.use_deterministic_algorithms instead. + warnings.warn( +You are using the default legacy behaviour of the . This is expected, and simply means that the `legacy` (previous) behavior will be used so nothing changes for you. If you want to use the new behaviour, set `legacy=False`. This should only be set if you understand what it means, and thoroughly read the reason why this was added as explained in https://github.com/huggingface/transformers/pull/24565 +2024-05-14:16:44:05,486 WARNING [task.py:763] [Task: indiccopa-hi] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-14:16:44:05,487 WARNING [task.py:775] [Task: indiccopa-hi] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +[2024-05-14 16:44:05,047] [INFO] [real_accelerator.py:178:get_accelerator] Setting ds_accelerator to hpu (auto detect) +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for ai4bharat/IndicCOPA contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/ai4bharat/IndicCOPA +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-05-14:16:44:06,789 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:16:44:06,789 WARNING [task.py:322] [Task: indiccopa-hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-05-14:16:44:06,808 INFO [task.py:395] Building contexts for indiccopa-hi on rank 1... +100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 56/56 [00:00<00:00, 104624.06it/s] +Passed argument batch_size = auto:1. Detecting largest batch size +2024-05-14:16:44:08,634 INFO [evaluator.py:379] Running loglikelihood requests +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Determined largest batch size: 64 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..c54937050fcef7bc8c9b8a7c7d7aa56fc5ab5130 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T16:43:51.649879", + "startedAt": "2024-05-14T16:43:51.257403", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3392.8983618421053, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 863.4305419921875 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..91df0012cef27fbd76437f2803da1fd4192acd69 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 24}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e010a0b22f9623386ab116d46929d6864dae4e19 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/logs/debug-internal.log @@ -0,0 +1,193 @@ +2024-05-14 16:43:51,270 INFO StreamThr :130068 [internal.py:wandb_internal():85] W&B internal server running at pid: 130068, started at: 2024-05-14 16:43:51.269256 +2024-05-14 16:43:51,272 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: status +2024-05-14 16:43:51,273 INFO WriterThread:130068 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/run-oynwi78z.wandb +2024-05-14 16:43:51,273 DEBUG SenderThread:130068 [sender.py:send():378] send: header +2024-05-14 16:43:51,285 DEBUG SenderThread:130068 [sender.py:send():378] send: run +2024-05-14 16:43:51,509 INFO SenderThread:130068 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files +2024-05-14 16:43:51,509 INFO SenderThread:130068 [sender.py:_start_run_threads():1123] run started: oynwi78z with start time 1715705031.268619 +2024-05-14 16:43:51,517 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 16:43:51,517 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: check_version +2024-05-14 16:43:51,601 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 16:43:51,603 DEBUG HandlerThread:130068 [system_info.py:__init__():26] System info init +2024-05-14 16:43:51,603 DEBUG HandlerThread:130068 [system_info.py:__init__():41] System info init done +2024-05-14 16:43:51,603 INFO HandlerThread:130068 [system_monitor.py:start():194] Starting system monitor +2024-05-14 16:43:51,603 INFO SystemMonitor:130068 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 16:43:51,603 INFO HandlerThread:130068 [system_monitor.py:probe():214] Collecting system info +2024-05-14 16:43:51,604 INFO SystemMonitor:130068 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 16:43:51,605 INFO SystemMonitor:130068 [interfaces.py:start():188] Started disk monitoring +2024-05-14 16:43:51,605 INFO SystemMonitor:130068 [interfaces.py:start():188] Started memory monitoring +2024-05-14 16:43:51,606 INFO SystemMonitor:130068 [interfaces.py:start():188] Started network monitoring +2024-05-14 16:43:51,649 DEBUG HandlerThread:130068 [system_info.py:probe():150] Probing system +2024-05-14 16:43:51,658 DEBUG HandlerThread:130068 [system_info.py:_probe_git():135] Probing git +2024-05-14 16:43:51,678 ERROR HandlerThread:130068 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 16:43:51,678 DEBUG HandlerThread:130068 [system_info.py:_probe_git():143] Probing git done +2024-05-14 16:43:51,678 DEBUG HandlerThread:130068 [system_info.py:probe():198] Probing system done +2024-05-14 16:43:51,678 DEBUG HandlerThread:130068 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T16:43:51.649879', 'startedAt': '2024-05-14T16:43:51.257403', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step100', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3392.8983618421053, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 863.4305419921875}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 16:43:51,679 INFO HandlerThread:130068 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 16:43:51,679 INFO HandlerThread:130068 [system_monitor.py:probe():227] Publishing system info +2024-05-14 16:43:51,680 INFO HandlerThread:130068 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 16:43:51,684 DEBUG SenderThread:130068 [sender.py:send():378] send: files +2024-05-14 16:43:51,684 INFO SenderThread:130068 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 16:43:51,780 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 16:43:51,780 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:43:51,781 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: python_packages +2024-05-14 16:43:51,781 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:43:51,976 DEBUG SenderThread:130068 [sender.py:send():378] send: telemetry +2024-05-14 16:43:52,223 INFO wandb-upload_0:130068 [upload_job.py:push():130] Uploaded file /tmp/tmpehqmmfb3wandb/uie0ldc9-wandb-metadata.json +2024-05-14 16:43:52,511 INFO Thread-12 :130068 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/requirements.txt +2024-05-14 16:43:52,512 INFO Thread-12 :130068 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:43:52,512 INFO Thread-12 :130068 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-metadata.json +2024-05-14 16:43:54,511 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:43:56,476 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:43:58,513 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:01,478 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:44:06,519 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:06,785 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:44:06,785 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:44:06,920 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:44:07,519 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:08,520 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:09,521 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:10,522 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:11,523 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:12,139 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:44:15,881 DEBUG SenderThread:130068 [sender.py:send():378] send: exit +2024-05-14 16:44:15,881 INFO SenderThread:130068 [sender.py:send_exit():585] handling exit code: 0 +2024-05-14 16:44:15,881 INFO SenderThread:130068 [sender.py:send_exit():587] handling runtime: 24 +2024-05-14 16:44:15,885 INFO SenderThread:130068 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:44:15,885 INFO SenderThread:130068 [sender.py:send_exit():593] send defer +2024-05-14 16:44:15,886 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:15,886 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 16:44:15,886 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:15,886 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 16:44:15,886 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 1 +2024-05-14 16:44:15,886 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:15,886 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 16:44:15,886 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:15,886 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 16:44:15,886 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 2 +2024-05-14 16:44:15,886 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:15,886 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 16:44:15,886 INFO HandlerThread:130068 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 16:44:15,886 DEBUG SystemMonitor:130068 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 16:44:15,887 DEBUG SystemMonitor:130068 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 16:44:15,887 DEBUG SystemMonitor:130068 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 16:44:15,887 INFO HandlerThread:130068 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 16:44:15,889 INFO HandlerThread:130068 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 16:44:15,889 INFO HandlerThread:130068 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 16:44:15,889 INFO HandlerThread:130068 [interfaces.py:finish():200] Joined network monitor +2024-05-14 16:44:15,889 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:15,889 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 16:44:15,889 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 3 +2024-05-14 16:44:15,890 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:15,890 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 16:44:15,890 DEBUG SenderThread:130068 [sender.py:send():378] send: stats +2024-05-14 16:44:15,890 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:15,890 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 16:44:15,891 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 4 +2024-05-14 16:44:15,891 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:15,891 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 16:44:15,891 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:15,891 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 16:44:15,891 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 5 +2024-05-14 16:44:15,891 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:15,891 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 16:44:15,891 DEBUG SenderThread:130068 [sender.py:send():378] send: summary +2024-05-14 16:44:15,892 INFO SenderThread:130068 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:44:15,892 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:15,892 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 16:44:15,892 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 6 +2024-05-14 16:44:15,892 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:15,892 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 16:44:15,892 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:15,892 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 16:44:15,895 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:44:15,985 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 7 +2024-05-14 16:44:15,985 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:15,985 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 16:44:15,985 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:15,985 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 16:44:16,527 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/config.yaml +2024-05-14 16:44:16,527 INFO Thread-12 :130068 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-summary.json +2024-05-14 16:44:16,881 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:44:17,062 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 8 +2024-05-14 16:44:17,062 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:44:17,062 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:17,062 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 16:44:17,062 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:17,063 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 16:44:17,063 INFO SenderThread:130068 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 16:44:17,063 INFO SenderThread:130068 [job_builder.py:_get_source_type():576] no source found +2024-05-14 16:44:17,063 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 9 +2024-05-14 16:44:17,063 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:17,063 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 16:44:17,063 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:17,063 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 16:44:17,063 INFO SenderThread:130068 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 16:44:17,528 INFO Thread-12 :130068 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:17,528 INFO SenderThread:130068 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files +2024-05-14 16:44:17,528 INFO SenderThread:130068 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/requirements.txt requirements.txt +2024-05-14 16:44:17,528 INFO SenderThread:130068 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-metadata.json wandb-metadata.json +2024-05-14 16:44:17,529 INFO SenderThread:130068 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/config.yaml config.yaml +2024-05-14 16:44:17,529 INFO SenderThread:130068 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log output.log +2024-05-14 16:44:17,529 INFO SenderThread:130068 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-summary.json wandb-summary.json +2024-05-14 16:44:17,529 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 10 +2024-05-14 16:44:17,529 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:17,529 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 16:44:17,532 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:17,534 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 16:44:17,534 INFO SenderThread:130068 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:44:17,764 INFO wandb-upload_1:130068 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/requirements.txt +2024-05-14 16:44:17,881 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:44:17,882 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:44:17,932 INFO wandb-upload_0:130068 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/config.yaml +2024-05-14 16:44:18,027 INFO wandb-upload_2:130068 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/output.log +2024-05-14 16:44:18,036 INFO wandb-upload_3:130068 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/files/wandb-summary.json +2024-05-14 16:44:18,237 INFO Thread-11 (_thread_body):130068 [sender.py:transition_state():613] send defer: 11 +2024-05-14 16:44:18,237 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:18,237 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 16:44:18,238 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:18,238 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 16:44:18,238 INFO SenderThread:130068 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 16:44:18,238 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 12 +2024-05-14 16:44:18,238 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:18,238 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 16:44:18,238 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:18,238 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 16:44:18,238 INFO SenderThread:130068 [file_stream.py:finish():601] file stream finish called +2024-05-14 16:44:18,300 INFO SenderThread:130068 [file_stream.py:finish():605] file stream finish is done +2024-05-14 16:44:18,300 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 13 +2024-05-14 16:44:18,300 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:18,300 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 16:44:18,301 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:18,301 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 16:44:18,301 INFO SenderThread:130068 [sender.py:transition_state():613] send defer: 14 +2024-05-14 16:44:18,301 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:44:18,301 INFO HandlerThread:130068 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 16:44:18,301 DEBUG SenderThread:130068 [sender.py:send():378] send: final +2024-05-14 16:44:18,301 DEBUG SenderThread:130068 [sender.py:send():378] send: footer +2024-05-14 16:44:18,301 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: defer +2024-05-14 16:44:18,301 INFO SenderThread:130068 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 16:44:18,302 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:44:18,302 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:44:18,302 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:44:18,302 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:44:18,302 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 16:44:18,303 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 16:44:18,303 DEBUG SenderThread:130068 [sender.py:send_request():405] send_request: server_info +2024-05-14 16:44:18,304 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 16:44:18,304 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 16:44:18,366 INFO MainThread:130068 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 16:44:18,366 INFO MainThread:130068 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 16:44:18,366 INFO MainThread:130068 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 16:44:18,366 DEBUG HandlerThread:130068 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 16:44:18,366 INFO HandlerThread:130068 [handler.py:finish():882] shutting down handler +2024-05-14 16:44:19,303 INFO WriterThread:130068 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/run-oynwi78z.wandb +2024-05-14 16:44:19,366 INFO SenderThread:130068 [sender.py:finish():1545] shutting down sender +2024-05-14 16:44:19,366 INFO SenderThread:130068 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:44:19,366 INFO SenderThread:130068 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..86b6c753665b1aa14adba3e2c40b435cb1e54956 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 16:43:51,265 INFO MainThread:128888 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 16:43:51,265 INFO MainThread:128888 [wandb_setup.py:_flush():76] Configure stats pid to 128888 +2024-05-14 16:43:51,265 INFO MainThread:128888 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 16:43:51,265 INFO MainThread:128888 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 16:43:51,265 INFO MainThread:128888 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 16:43:51,265 INFO MainThread:128888 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 16:43:51,265 WARNING MainThread:128888 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 16:43:51,265 INFO MainThread:128888 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 16:43:51,265 INFO MainThread:128888 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 16:43:51,266 INFO MainThread:128888 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/logs/debug.log +2024-05-14 16:43:51,266 INFO MainThread:128888 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/logs/debug-internal.log +2024-05-14 16:43:51,266 INFO MainThread:128888 [wandb_init.py:init():560] calling init triggers +2024-05-14 16:43:51,266 INFO MainThread:128888 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 16:43:51,266 INFO MainThread:128888 [wandb_init.py:init():610] starting backend +2024-05-14 16:43:51,266 INFO MainThread:128888 [wandb_init.py:init():614] setting up manager +2024-05-14 16:43:51,267 INFO MainThread:128888 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 16:43:51,268 INFO MainThread:128888 [wandb_init.py:init():622] backend started and connected +2024-05-14 16:43:51,271 INFO MainThread:128888 [wandb_init.py:init():711] updated telemetry +2024-05-14 16:43:51,284 INFO MainThread:128888 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 16:43:51,516 INFO MainThread:128888 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 16:43:51,597 INFO MainThread:128888 [wandb_run.py:_on_init():2405] got version response +2024-05-14 16:43:51,597 INFO MainThread:128888 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 16:43:51,781 INFO MainThread:128888 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 16:43:51,781 INFO MainThread:128888 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 16:43:51,781 INFO MainThread:128888 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 16:43:51,781 INFO MainThread:128888 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 16:43:51,782 INFO MainThread:128888 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 16:44:19,367 WARNING MsgRouterThr:128888 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/run-oynwi78z.wandb b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/run-oynwi78z.wandb new file mode 100644 index 0000000000000000000000000000000000000000..2e20d8933a83a167cb1b1aeb2369b13921773485 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_164351-oynwi78z/run-oynwi78z.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..46cf6f00b3bd7716effe94e46d60c979920d0178 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715705121 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/output.log b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..81b791b5f30221839fe40e9024726225b7f46841 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/output.log @@ -0,0 +1,33 @@ + +2024-05-14:16:45:22,431 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:16:45:26,990 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:16:45:26,993 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:16:45:26,993 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step120'} +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/data/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/data/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/data/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 928, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 631, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 686, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 369, in cached_file + raise EnvironmentError( +OSError: /data/cronscript/ckpts//hf_ckpt//global_step120 does not appear to have a file named config.json. Checkout 'https://huggingface.co//data/cronscript/ckpts//hf_ckpt//global_step120/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..58b2ae765fb7524e08b57a2cead9f4fd026544e8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T16:45:22.298330", + "startedAt": "2024-05-14T16:45:21.857942", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3390.1116644736844, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.888, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3271.761, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3284.449, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.005, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.002, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 863.430061340332 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 5}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..32116dc550e2d630c7ebf96ab5d2e6d29d6ae945 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/logs/debug-internal.log @@ -0,0 +1,179 @@ +2024-05-14 16:45:21,869 INFO StreamThr :137239 [internal.py:wandb_internal():85] W&B internal server running at pid: 137239, started at: 2024-05-14 16:45:21.868632 +2024-05-14 16:45:21,871 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: status +2024-05-14 16:45:21,872 INFO WriterThread:137239 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/run-i74dki84.wandb +2024-05-14 16:45:21,878 DEBUG SenderThread:137239 [sender.py:send():378] send: header +2024-05-14 16:45:21,890 DEBUG SenderThread:137239 [sender.py:send():378] send: run +2024-05-14 16:45:22,137 INFO SenderThread:137239 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files +2024-05-14 16:45:22,137 INFO SenderThread:137239 [sender.py:_start_run_threads():1123] run started: i74dki84 with start time 1715705121.868767 +2024-05-14 16:45:22,145 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 16:45:22,145 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: check_version +2024-05-14 16:45:22,234 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 16:45:22,236 DEBUG HandlerThread:137239 [system_info.py:__init__():26] System info init +2024-05-14 16:45:22,236 DEBUG HandlerThread:137239 [system_info.py:__init__():41] System info init done +2024-05-14 16:45:22,236 INFO HandlerThread:137239 [system_monitor.py:start():194] Starting system monitor +2024-05-14 16:45:22,236 INFO SystemMonitor:137239 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 16:45:22,236 INFO HandlerThread:137239 [system_monitor.py:probe():214] Collecting system info +2024-05-14 16:45:22,236 INFO SystemMonitor:137239 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 16:45:22,237 INFO SystemMonitor:137239 [interfaces.py:start():188] Started disk monitoring +2024-05-14 16:45:22,237 INFO SystemMonitor:137239 [interfaces.py:start():188] Started memory monitoring +2024-05-14 16:45:22,237 INFO SystemMonitor:137239 [interfaces.py:start():188] Started network monitoring +2024-05-14 16:45:22,298 DEBUG HandlerThread:137239 [system_info.py:probe():150] Probing system +2024-05-14 16:45:22,306 DEBUG HandlerThread:137239 [system_info.py:_probe_git():135] Probing git +2024-05-14 16:45:22,327 ERROR HandlerThread:137239 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 16:45:22,327 DEBUG HandlerThread:137239 [system_info.py:_probe_git():143] Probing git done +2024-05-14 16:45:22,327 DEBUG HandlerThread:137239 [system_info.py:probe():198] Probing system done +2024-05-14 16:45:22,328 DEBUG HandlerThread:137239 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T16:45:22.298330', 'startedAt': '2024-05-14T16:45:21.857942', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3390.1116644736844, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.888, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3271.761, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3284.449, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.005, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.002, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.002, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 863.430061340332}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 16:45:22,328 INFO HandlerThread:137239 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 16:45:22,328 INFO HandlerThread:137239 [system_monitor.py:probe():227] Publishing system info +2024-05-14 16:45:22,329 INFO HandlerThread:137239 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 16:45:22,332 DEBUG SenderThread:137239 [sender.py:send():378] send: files +2024-05-14 16:45:22,333 INFO SenderThread:137239 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 16:45:22,428 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 16:45:22,429 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:45:22,429 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: python_packages +2024-05-14 16:45:22,430 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:45:22,603 DEBUG SenderThread:137239 [sender.py:send():378] send: telemetry +2024-05-14 16:45:22,847 INFO wandb-upload_0:137239 [upload_job.py:push():130] Uploaded file /tmp/tmpi5klit_6wandb/xkmgv4pt-wandb-metadata.json +2024-05-14 16:45:23,139 INFO Thread-12 :137239 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-metadata.json +2024-05-14 16:45:23,139 INFO Thread-12 :137239 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/requirements.txt +2024-05-14 16:45:23,139 INFO Thread-12 :137239 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/output.log +2024-05-14 16:45:25,140 INFO Thread-12 :137239 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/output.log +2024-05-14 16:45:26,991 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:45:28,083 DEBUG SenderThread:137239 [sender.py:send():378] send: exit +2024-05-14 16:45:28,083 INFO SenderThread:137239 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 16:45:28,084 INFO SenderThread:137239 [sender.py:send_exit():587] handling runtime: 5 +2024-05-14 16:45:28,085 INFO SenderThread:137239 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:45:28,085 INFO SenderThread:137239 [sender.py:send_exit():593] send defer +2024-05-14 16:45:28,085 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,085 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 16:45:28,085 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,085 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 16:45:28,085 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 1 +2024-05-14 16:45:28,085 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,085 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 16:45:28,085 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,085 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 16:45:28,085 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 2 +2024-05-14 16:45:28,085 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,085 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 16:45:28,086 INFO HandlerThread:137239 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 16:45:28,086 DEBUG SystemMonitor:137239 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 16:45:28,086 INFO HandlerThread:137239 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 16:45:28,086 DEBUG SystemMonitor:137239 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 16:45:28,086 INFO HandlerThread:137239 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 16:45:28,086 DEBUG SystemMonitor:137239 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 16:45:28,086 INFO HandlerThread:137239 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 16:45:28,088 INFO HandlerThread:137239 [interfaces.py:finish():200] Joined network monitor +2024-05-14 16:45:28,088 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,088 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 16:45:28,088 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 3 +2024-05-14 16:45:28,088 DEBUG SenderThread:137239 [sender.py:send():378] send: stats +2024-05-14 16:45:28,088 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,089 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 16:45:28,089 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,089 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 16:45:28,089 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 4 +2024-05-14 16:45:28,089 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,089 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 16:45:28,089 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,089 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 16:45:28,089 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 5 +2024-05-14 16:45:28,089 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,089 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 16:45:28,089 DEBUG SenderThread:137239 [sender.py:send():378] send: summary +2024-05-14 16:45:28,090 INFO SenderThread:137239 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:45:28,090 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,090 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 16:45:28,090 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 6 +2024-05-14 16:45:28,090 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,090 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 16:45:28,090 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,090 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 16:45:28,093 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:45:28,141 INFO Thread-12 :137239 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-summary.json +2024-05-14 16:45:28,165 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 7 +2024-05-14 16:45:28,165 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,165 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 16:45:28,165 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,165 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 16:45:28,620 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 8 +2024-05-14 16:45:28,620 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,620 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 16:45:28,621 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,621 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 16:45:28,621 INFO SenderThread:137239 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 16:45:28,621 INFO SenderThread:137239 [job_builder.py:_get_source_type():576] no source found +2024-05-14 16:45:28,621 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 9 +2024-05-14 16:45:28,621 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,621 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 16:45:28,621 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,622 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 16:45:28,622 INFO SenderThread:137239 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 16:45:29,084 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:29,142 INFO SenderThread:137239 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/config.yaml +2024-05-14 16:45:29,142 INFO SenderThread:137239 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/output.log +2024-05-14 16:45:29,143 INFO SenderThread:137239 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files +2024-05-14 16:45:29,143 INFO SenderThread:137239 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-metadata.json wandb-metadata.json +2024-05-14 16:45:29,143 INFO SenderThread:137239 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/output.log output.log +2024-05-14 16:45:29,143 INFO SenderThread:137239 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-summary.json wandb-summary.json +2024-05-14 16:45:29,143 INFO SenderThread:137239 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/requirements.txt requirements.txt +2024-05-14 16:45:29,143 INFO SenderThread:137239 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/config.yaml config.yaml +2024-05-14 16:45:29,143 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 10 +2024-05-14 16:45:29,143 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:29,143 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:29,146 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 16:45:29,149 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:29,149 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 16:45:29,149 INFO SenderThread:137239 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:45:29,383 INFO wandb-upload_2:137239 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/requirements.txt +2024-05-14 16:45:29,537 INFO wandb-upload_0:137239 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/output.log +2024-05-14 16:45:29,603 INFO wandb-upload_3:137239 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/config.yaml +2024-05-14 16:45:29,614 INFO wandb-upload_1:137239 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/files/wandb-summary.json +2024-05-14 16:45:29,815 INFO Thread-11 (_thread_body):137239 [sender.py:transition_state():613] send defer: 11 +2024-05-14 16:45:29,815 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:29,815 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 16:45:29,816 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:29,816 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 16:45:29,816 INFO SenderThread:137239 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 16:45:29,816 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 12 +2024-05-14 16:45:29,816 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:29,816 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 16:45:29,816 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:29,816 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 16:45:29,816 INFO SenderThread:137239 [file_stream.py:finish():601] file stream finish called +2024-05-14 16:45:30,028 INFO SenderThread:137239 [file_stream.py:finish():605] file stream finish is done +2024-05-14 16:45:30,028 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 13 +2024-05-14 16:45:30,028 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:30,028 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 16:45:30,028 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:30,028 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 16:45:30,029 INFO SenderThread:137239 [sender.py:transition_state():613] send defer: 14 +2024-05-14 16:45:30,029 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:30,029 INFO HandlerThread:137239 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 16:45:30,029 DEBUG SenderThread:137239 [sender.py:send():378] send: final +2024-05-14 16:45:30,029 DEBUG SenderThread:137239 [sender.py:send():378] send: footer +2024-05-14 16:45:30,029 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:30,029 INFO SenderThread:137239 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 16:45:30,029 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:30,029 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:30,030 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:30,030 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:30,030 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 16:45:30,030 DEBUG SenderThread:137239 [sender.py:send_request():405] send_request: server_info +2024-05-14 16:45:30,030 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 16:45:30,031 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 16:45:30,031 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 16:45:30,082 INFO MainThread:137239 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 16:45:30,082 INFO MainThread:137239 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 16:45:30,082 INFO MainThread:137239 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 16:45:30,082 DEBUG HandlerThread:137239 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 16:45:30,083 INFO HandlerThread:137239 [handler.py:finish():882] shutting down handler +2024-05-14 16:45:31,030 INFO WriterThread:137239 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/run-i74dki84.wandb +2024-05-14 16:45:31,082 INFO SenderThread:137239 [sender.py:finish():1545] shutting down sender +2024-05-14 16:45:31,082 INFO SenderThread:137239 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:45:31,082 INFO SenderThread:137239 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..56f93462a6065f69886093f5cf239d71fd94a67a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 16:45:21,865 INFO MainThread:136018 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 16:45:21,865 INFO MainThread:136018 [wandb_setup.py:_flush():76] Configure stats pid to 136018 +2024-05-14 16:45:21,865 INFO MainThread:136018 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 16:45:21,866 WARNING MainThread:136018 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/logs/debug.log +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/logs/debug-internal.log +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_init.py:init():560] calling init triggers +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_init.py:init():610] starting backend +2024-05-14 16:45:21,866 INFO MainThread:136018 [wandb_init.py:init():614] setting up manager +2024-05-14 16:45:21,867 INFO MainThread:136018 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 16:45:21,868 INFO MainThread:136018 [wandb_init.py:init():622] backend started and connected +2024-05-14 16:45:21,872 INFO MainThread:136018 [wandb_init.py:init():711] updated telemetry +2024-05-14 16:45:21,889 INFO MainThread:136018 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 16:45:22,144 INFO MainThread:136018 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 16:45:22,229 INFO MainThread:136018 [wandb_run.py:_on_init():2405] got version response +2024-05-14 16:45:22,229 INFO MainThread:136018 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 16:45:22,429 INFO MainThread:136018 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 16:45:22,429 INFO MainThread:136018 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 16:45:22,429 INFO MainThread:136018 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 16:45:22,429 INFO MainThread:136018 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 16:45:22,430 INFO MainThread:136018 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 16:45:31,083 WARNING MsgRouterThr:136018 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/run-i74dki84.wandb b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/run-i74dki84.wandb new file mode 100644 index 0000000000000000000000000000000000000000..71864c8e6353d5f1b7cc27e25fcc4e2d45d91ae8 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_164521-i74dki84/run-i74dki84.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..d246150eb7e2a070b664d5e86610e1b2eb0f9d32 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716396318 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..876c9292e51535ab61336dc872155f79a57a8cde --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log @@ -0,0 +1,43 @@ + +2024-05-22:16:45:19,145 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:16:45:27,633 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:16:45:27,633 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:16:45:27,634 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000'} +2024-05-22:16:45:29,947 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 399, in cached_file + resolved_file = hf_hub_download( + File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py", line 106, in _inner_fn + validate_repo_id(arg_value) + File "/usr/local/lib/python3.10/dist-packages/huggingface_hub/utils/_validators.py", line 154, in validate_repo_id + raise HFValidationError( +huggingface_hub.errors.HFValidationError: Repo id must be in the form 'repo_name' or 'namespace/repo_name': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000'. Use `repo_type` argument if needed. +The above exception was the direct cause of the following exception: +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 463, in cached_file + raise EnvironmentError( +OSError: Incorrect path_or_model_id: '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000'. Please provide either the path to a local folder or the repo_id of a model on the Hub. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..e3a54da98d28b517343fd8d6d3c350f9b88eb14a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T16:45:18.941515", + "startedAt": "2024-05-22T16:45:18.361820", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2334.215225, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64148330688477 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..4dbd61f3e1a893589ce13c70ce86a304a000f1a1 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/logs/debug-internal.log @@ -0,0 +1,182 @@ +2024-05-22 16:45:18,382 INFO StreamThr :1284 [internal.py:wandb_internal():85] W&B internal server running at pid: 1284, started at: 2024-05-22 16:45:18.381070 +2024-05-22 16:45:18,387 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: status +2024-05-22 16:45:18,388 INFO WriterThread:1284 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/run-ek2h8q84.wandb +2024-05-22 16:45:18,390 DEBUG SenderThread:1284 [sender.py:send():378] send: header +2024-05-22 16:45:18,394 DEBUG SenderThread:1284 [sender.py:send():378] send: run +2024-05-22 16:45:18,747 INFO SenderThread:1284 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files +2024-05-22 16:45:18,747 INFO SenderThread:1284 [sender.py:_start_run_threads():1123] run started: ek2h8q84 with start time 1716396318.380929 +2024-05-22 16:45:18,752 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 16:45:18,752 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: check_version +2024-05-22 16:45:18,866 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 16:45:18,869 DEBUG HandlerThread:1284 [system_info.py:__init__():26] System info init +2024-05-22 16:45:18,869 DEBUG HandlerThread:1284 [system_info.py:__init__():41] System info init done +2024-05-22 16:45:18,869 INFO HandlerThread:1284 [system_monitor.py:start():194] Starting system monitor +2024-05-22 16:45:18,869 INFO SystemMonitor:1284 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 16:45:18,869 INFO HandlerThread:1284 [system_monitor.py:probe():214] Collecting system info +2024-05-22 16:45:18,876 INFO SystemMonitor:1284 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 16:45:18,876 INFO SystemMonitor:1284 [interfaces.py:start():188] Started disk monitoring +2024-05-22 16:45:18,879 INFO SystemMonitor:1284 [interfaces.py:start():188] Started memory monitoring +2024-05-22 16:45:18,879 INFO SystemMonitor:1284 [interfaces.py:start():188] Started network monitoring +2024-05-22 16:45:18,941 DEBUG HandlerThread:1284 [system_info.py:probe():150] Probing system +2024-05-22 16:45:18,944 DEBUG HandlerThread:1284 [system_info.py:_probe_git():135] Probing git +2024-05-22 16:45:18,955 ERROR HandlerThread:1284 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 16:45:18,955 DEBUG HandlerThread:1284 [system_info.py:_probe_git():143] Probing git done +2024-05-22 16:45:18,955 DEBUG HandlerThread:1284 [system_info.py:probe():198] Probing system done +2024-05-22 16:45:18,955 DEBUG HandlerThread:1284 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T16:45:18.941515', 'startedAt': '2024-05-22T16:45:18.361820', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2334.215225, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64148330688477}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 16:45:18,955 INFO HandlerThread:1284 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 16:45:18,955 INFO HandlerThread:1284 [system_monitor.py:probe():227] Publishing system info +2024-05-22 16:45:18,958 INFO HandlerThread:1284 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 16:45:18,963 DEBUG SenderThread:1284 [sender.py:send():378] send: files +2024-05-22 16:45:18,963 INFO SenderThread:1284 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 16:45:19,138 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 16:45:19,138 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: python_packages +2024-05-22 16:45:19,139 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 16:45:19,141 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: stop_status +2024-05-22 16:45:19,212 DEBUG SenderThread:1284 [sender.py:send():378] send: telemetry +2024-05-22 16:45:19,516 INFO wandb-upload_0:1284 [upload_job.py:push():130] Uploaded file /tmp/tmp0ri0ctcdwandb/wrkg47rc-wandb-metadata.json +2024-05-22 16:45:19,750 INFO Thread-12 :1284 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-metadata.json +2024-05-22 16:45:19,750 INFO Thread-12 :1284 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/requirements.txt +2024-05-22 16:45:19,750 INFO Thread-12 :1284 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log +2024-05-22 16:45:21,750 INFO Thread-12 :1284 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log +2024-05-22 16:45:24,215 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 16:45:29,634 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 16:45:29,756 INFO Thread-12 :1284 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log +2024-05-22 16:45:29,956 DEBUG SenderThread:1284 [sender.py:send():378] send: exit +2024-05-22 16:45:29,956 INFO SenderThread:1284 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 16:45:29,956 INFO SenderThread:1284 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 16:45:29,958 INFO SenderThread:1284 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 16:45:29,958 INFO SenderThread:1284 [sender.py:send_exit():593] send defer +2024-05-22 16:45:29,958 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:29,958 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 16:45:29,958 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:29,958 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 16:45:29,958 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 1 +2024-05-22 16:45:29,958 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:29,958 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 16:45:29,958 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:29,958 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 16:45:29,958 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 2 +2024-05-22 16:45:29,959 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:29,959 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 16:45:29,959 INFO HandlerThread:1284 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 16:45:29,959 DEBUG SystemMonitor:1284 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 16:45:29,959 DEBUG SystemMonitor:1284 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 16:45:29,959 DEBUG SystemMonitor:1284 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 16:45:29,960 INFO HandlerThread:1284 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 16:45:29,960 INFO HandlerThread:1284 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 16:45:29,960 INFO HandlerThread:1284 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 16:45:29,960 INFO HandlerThread:1284 [interfaces.py:finish():200] Joined network monitor +2024-05-22 16:45:29,960 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:29,960 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 16:45:29,960 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 3 +2024-05-22 16:45:29,960 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:29,960 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 16:45:29,960 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:29,960 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 16:45:29,961 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 4 +2024-05-22 16:45:29,961 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:29,961 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 16:45:29,961 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:29,961 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 16:45:29,961 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 5 +2024-05-22 16:45:29,961 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:29,961 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 16:45:29,961 DEBUG SenderThread:1284 [sender.py:send():378] send: summary +2024-05-22 16:45:29,962 INFO SenderThread:1284 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 16:45:29,962 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:29,962 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 16:45:29,962 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 6 +2024-05-22 16:45:29,962 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:29,962 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 16:45:29,962 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:29,963 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 16:45:29,967 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 16:45:30,033 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 7 +2024-05-22 16:45:30,033 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:30,033 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 16:45:30,034 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:30,034 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 16:45:30,758 INFO Thread-12 :1284 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/config.yaml +2024-05-22 16:45:30,758 INFO Thread-12 :1284 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-summary.json +2024-05-22 16:45:30,956 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 16:45:31,239 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 8 +2024-05-22 16:45:31,239 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 16:45:31,239 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:31,239 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 16:45:31,240 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:31,240 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 16:45:31,240 INFO SenderThread:1284 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 16:45:31,240 INFO SenderThread:1284 [job_builder.py:_get_source_type():576] no source found +2024-05-22 16:45:31,240 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 9 +2024-05-22 16:45:31,241 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:31,241 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 16:45:31,241 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:31,241 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 16:45:31,241 INFO SenderThread:1284 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 16:45:31,759 INFO SenderThread:1284 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log +2024-05-22 16:45:31,760 INFO SenderThread:1284 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files +2024-05-22 16:45:31,760 INFO SenderThread:1284 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-summary.json wandb-summary.json +2024-05-22 16:45:31,760 INFO SenderThread:1284 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/config.yaml config.yaml +2024-05-22 16:45:31,762 INFO SenderThread:1284 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/requirements.txt requirements.txt +2024-05-22 16:45:31,764 INFO SenderThread:1284 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-metadata.json wandb-metadata.json +2024-05-22 16:45:31,765 INFO SenderThread:1284 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log output.log +2024-05-22 16:45:31,765 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 10 +2024-05-22 16:45:31,767 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:31,767 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 16:45:31,767 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:31,767 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 16:45:31,767 INFO SenderThread:1284 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 16:45:31,956 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 16:45:31,956 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 16:45:32,079 INFO wandb-upload_0:1284 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/wandb-summary.json +2024-05-22 16:45:32,363 INFO wandb-upload_3:1284 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/output.log +2024-05-22 16:45:32,371 INFO wandb-upload_2:1284 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/requirements.txt +2024-05-22 16:45:32,378 INFO wandb-upload_1:1284 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/files/config.yaml +2024-05-22 16:45:32,578 INFO Thread-11 (_thread_body):1284 [sender.py:transition_state():613] send defer: 11 +2024-05-22 16:45:32,579 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:32,579 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 16:45:32,579 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:32,579 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 16:45:32,579 INFO SenderThread:1284 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 16:45:32,579 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 12 +2024-05-22 16:45:32,579 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:32,579 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 16:45:32,579 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:32,579 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 16:45:32,579 INFO SenderThread:1284 [file_stream.py:finish():601] file stream finish called +2024-05-22 16:45:32,654 INFO SenderThread:1284 [file_stream.py:finish():605] file stream finish is done +2024-05-22 16:45:32,654 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 13 +2024-05-22 16:45:32,654 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:32,654 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 16:45:32,654 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:32,654 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 16:45:32,654 INFO SenderThread:1284 [sender.py:transition_state():613] send defer: 14 +2024-05-22 16:45:32,654 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:45:32,654 INFO HandlerThread:1284 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 16:45:32,655 DEBUG SenderThread:1284 [sender.py:send():378] send: final +2024-05-22 16:45:32,655 DEBUG SenderThread:1284 [sender.py:send():378] send: footer +2024-05-22 16:45:32,655 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: defer +2024-05-22 16:45:32,655 INFO SenderThread:1284 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 16:45:32,655 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 16:45:32,656 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 16:45:32,656 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 16:45:32,656 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 16:45:32,656 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 16:45:32,656 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 16:45:32,656 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 16:45:32,656 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 16:45:32,656 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: server_info +2024-05-22 16:45:32,731 INFO MainThread:1284 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 16:45:32,731 INFO MainThread:1284 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 16:45:32,731 INFO MainThread:1284 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-22 16:45:32,731 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 16:45:32,731 INFO HandlerThread:1284 [handler.py:finish():882] shutting down handler +2024-05-22 16:45:33,656 INFO WriterThread:1284 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/run-ek2h8q84.wandb +2024-05-22 16:45:33,731 INFO SenderThread:1284 [sender.py:finish():1545] shutting down sender +2024-05-22 16:45:33,731 INFO SenderThread:1284 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 16:45:33,731 INFO SenderThread:1284 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..6a978c6bd0439e9fa7b53bfd336ab57bc0f53785 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_setup.py:_flush():76] Configure stats pid to 1129 +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 16:45:18,375 WARNING MainThread:1129 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/logs/debug.log +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164518-ek2h8q84/logs/debug-internal.log +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_init.py:init():560] calling init triggers +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_init.py:init():610] starting backend +2024-05-22 16:45:18,375 INFO MainThread:1129 [wandb_init.py:init():614] setting up manager +2024-05-22 16:45:18,379 INFO MainThread:1129 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 16:45:18,380 INFO MainThread:1129 [wandb_init.py:init():622] backend started and connected +2024-05-22 16:45:18,384 INFO MainThread:1129 [wandb_init.py:init():711] updated telemetry +2024-05-22 16:45:18,393 INFO MainThread:1129 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 16:45:18,751 INFO MainThread:1129 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 16:45:18,860 INFO MainThread:1129 [wandb_run.py:_on_init():2405] got version response +2024-05-22 16:45:18,861 INFO MainThread:1129 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 16:45:19,139 INFO MainThread:1129 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 16:45:19,139 INFO MainThread:1129 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 16:45:19,139 INFO MainThread:1129 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 16:45:19,139 INFO MainThread:1129 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 16:45:19,143 INFO MainThread:1129 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 16:45:33,732 WARNING MsgRouterThr:1129 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..eb1c28c2007f47613dcdf60d46b0a1498bfdc690 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716404313 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..984df1939c756a093492a2c466ce71abe7867555 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log @@ -0,0 +1,34 @@ + +2024-05-22:18:58:34,396 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:18:58:42,864 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:18:58:42,865 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:18:58:42,866 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step28000'} +2024-05-22:18:58:45,166 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step28000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step28000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..69729ce1ea24874027abf6796109b92a296bf94f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T18:58:34.197313", + "startedAt": "2024-05-22T18:58:33.638991", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step28000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.5000124999997, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64125442504883 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..4cbfc41c1d548da42175d65d0cf1789e6f400f01 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-22 18:58:33,663 INFO StreamThr :3594 [internal.py:wandb_internal():85] W&B internal server running at pid: 3594, started at: 2024-05-22 18:58:33.661269 +2024-05-22 18:58:33,664 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: status +2024-05-22 18:58:33,666 INFO WriterThread:3594 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/run-auaa3wb6.wandb +2024-05-22 18:58:33,670 DEBUG SenderThread:3594 [sender.py:send():378] send: header +2024-05-22 18:58:33,675 DEBUG SenderThread:3594 [sender.py:send():378] send: run +2024-05-22 18:58:33,975 INFO SenderThread:3594 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files +2024-05-22 18:58:33,975 INFO SenderThread:3594 [sender.py:_start_run_threads():1123] run started: auaa3wb6 with start time 1716404313.661745 +2024-05-22 18:58:33,979 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 18:58:33,979 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: check_version +2024-05-22 18:58:34,103 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 18:58:34,106 DEBUG HandlerThread:3594 [system_info.py:__init__():26] System info init +2024-05-22 18:58:34,106 DEBUG HandlerThread:3594 [system_info.py:__init__():41] System info init done +2024-05-22 18:58:34,106 INFO HandlerThread:3594 [system_monitor.py:start():194] Starting system monitor +2024-05-22 18:58:34,106 INFO SystemMonitor:3594 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 18:58:34,106 INFO HandlerThread:3594 [system_monitor.py:probe():214] Collecting system info +2024-05-22 18:58:34,113 INFO SystemMonitor:3594 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 18:58:34,113 INFO SystemMonitor:3594 [interfaces.py:start():188] Started disk monitoring +2024-05-22 18:58:34,116 INFO SystemMonitor:3594 [interfaces.py:start():188] Started memory monitoring +2024-05-22 18:58:34,121 INFO SystemMonitor:3594 [interfaces.py:start():188] Started network monitoring +2024-05-22 18:58:34,197 DEBUG HandlerThread:3594 [system_info.py:probe():150] Probing system +2024-05-22 18:58:34,200 DEBUG HandlerThread:3594 [system_info.py:_probe_git():135] Probing git +2024-05-22 18:58:34,210 ERROR HandlerThread:3594 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 18:58:34,210 DEBUG HandlerThread:3594 [system_info.py:_probe_git():143] Probing git done +2024-05-22 18:58:34,210 DEBUG HandlerThread:3594 [system_info.py:probe():198] Probing system done +2024-05-22 18:58:34,211 DEBUG HandlerThread:3594 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T18:58:34.197313', 'startedAt': '2024-05-22T18:58:33.638991', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step28000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.5000124999997, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64125442504883}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 18:58:34,211 INFO HandlerThread:3594 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 18:58:34,211 INFO HandlerThread:3594 [system_monitor.py:probe():227] Publishing system info +2024-05-22 18:58:34,214 INFO HandlerThread:3594 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 18:58:34,218 DEBUG SenderThread:3594 [sender.py:send():378] send: files +2024-05-22 18:58:34,219 INFO SenderThread:3594 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 18:58:34,390 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 18:58:34,390 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: python_packages +2024-05-22 18:58:34,392 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 18:58:34,392 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: stop_status +2024-05-22 18:58:34,567 DEBUG SenderThread:3594 [sender.py:send():378] send: telemetry +2024-05-22 18:58:34,763 INFO wandb-upload_0:3594 [upload_job.py:push():130] Uploaded file /tmp/tmp7n7gc4_wwandb/0ulqgjg0-wandb-metadata.json +2024-05-22 18:58:34,977 INFO Thread-12 :3594 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/requirements.txt +2024-05-22 18:58:34,978 INFO Thread-12 :3594 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log +2024-05-22 18:58:34,978 INFO Thread-12 :3594 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-metadata.json +2024-05-22 18:58:36,977 INFO Thread-12 :3594 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log +2024-05-22 18:58:39,570 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:58:44,866 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:58:44,989 INFO Thread-12 :3594 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log +2024-05-22 18:58:45,173 DEBUG SenderThread:3594 [sender.py:send():378] send: exit +2024-05-22 18:58:45,173 INFO SenderThread:3594 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 18:58:45,173 INFO SenderThread:3594 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 18:58:45,179 INFO SenderThread:3594 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:58:45,179 INFO SenderThread:3594 [sender.py:send_exit():593] send defer +2024-05-22 18:58:45,179 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:45,179 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 18:58:45,179 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:45,179 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 18:58:45,179 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 1 +2024-05-22 18:58:45,179 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:45,180 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 18:58:45,180 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:45,180 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 18:58:45,180 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 2 +2024-05-22 18:58:45,180 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:45,180 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 18:58:45,180 INFO HandlerThread:3594 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 18:58:45,180 DEBUG SystemMonitor:3594 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 18:58:45,181 DEBUG SystemMonitor:3594 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 18:58:45,181 INFO HandlerThread:3594 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 18:58:45,181 DEBUG SystemMonitor:3594 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 18:58:45,181 INFO HandlerThread:3594 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 18:58:45,182 INFO HandlerThread:3594 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 18:58:45,183 INFO HandlerThread:3594 [interfaces.py:finish():200] Joined network monitor +2024-05-22 18:58:45,183 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:45,183 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 18:58:45,183 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 3 +2024-05-22 18:58:45,183 DEBUG SenderThread:3594 [sender.py:send():378] send: stats +2024-05-22 18:58:45,184 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:45,184 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 18:58:45,184 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:45,184 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 18:58:45,184 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 4 +2024-05-22 18:58:45,185 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:45,185 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 18:58:45,185 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:45,185 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 18:58:45,185 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 5 +2024-05-22 18:58:45,185 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:45,185 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 18:58:45,185 DEBUG SenderThread:3594 [sender.py:send():378] send: summary +2024-05-22 18:58:45,186 INFO SenderThread:3594 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:58:45,186 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:45,186 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 18:58:45,186 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 6 +2024-05-22 18:58:45,186 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:45,186 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 18:58:45,186 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:45,186 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 18:58:45,191 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:58:45,258 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 7 +2024-05-22 18:58:45,258 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:45,258 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 18:58:45,258 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:45,258 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 18:58:45,990 INFO Thread-12 :3594 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/config.yaml +2024-05-22 18:58:45,991 INFO Thread-12 :3594 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-summary.json +2024-05-22 18:58:46,173 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:58:46,590 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 8 +2024-05-22 18:58:46,590 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:58:46,590 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:46,591 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 18:58:46,591 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:46,591 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 18:58:46,591 INFO SenderThread:3594 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 18:58:46,591 INFO SenderThread:3594 [job_builder.py:_get_source_type():576] no source found +2024-05-22 18:58:46,592 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 9 +2024-05-22 18:58:46,592 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:46,592 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 18:58:46,592 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:46,592 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 18:58:46,592 INFO SenderThread:3594 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 18:58:46,992 INFO SenderThread:3594 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log +2024-05-22 18:58:46,992 INFO SenderThread:3594 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files +2024-05-22 18:58:46,992 INFO SenderThread:3594 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-metadata.json wandb-metadata.json +2024-05-22 18:58:46,992 INFO SenderThread:3594 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/config.yaml config.yaml +2024-05-22 18:58:46,992 INFO SenderThread:3594 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-summary.json wandb-summary.json +2024-05-22 18:58:46,995 INFO SenderThread:3594 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/requirements.txt requirements.txt +2024-05-22 18:58:46,997 INFO SenderThread:3594 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log output.log +2024-05-22 18:58:46,997 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 10 +2024-05-22 18:58:46,997 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:46,997 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 18:58:47,000 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:47,000 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 18:58:47,000 INFO SenderThread:3594 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:58:47,173 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:58:47,174 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:58:47,245 INFO wandb-upload_0:3594 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/config.yaml +2024-05-22 18:58:47,563 INFO wandb-upload_1:3594 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/wandb-summary.json +2024-05-22 18:58:47,598 INFO wandb-upload_2:3594 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/requirements.txt +2024-05-22 18:58:47,603 INFO wandb-upload_3:3594 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/files/output.log +2024-05-22 18:58:47,803 INFO Thread-11 (_thread_body):3594 [sender.py:transition_state():613] send defer: 11 +2024-05-22 18:58:47,803 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:47,804 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 18:58:47,804 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:47,804 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 18:58:47,804 INFO SenderThread:3594 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 18:58:47,804 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 12 +2024-05-22 18:58:47,804 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:47,804 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 18:58:47,804 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:47,804 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 18:58:47,804 INFO SenderThread:3594 [file_stream.py:finish():601] file stream finish called +2024-05-22 18:58:47,865 INFO SenderThread:3594 [file_stream.py:finish():605] file stream finish is done +2024-05-22 18:58:47,866 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 13 +2024-05-22 18:58:47,866 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:47,866 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 18:58:47,866 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:47,866 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 18:58:47,866 INFO SenderThread:3594 [sender.py:transition_state():613] send defer: 14 +2024-05-22 18:58:47,866 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:58:47,866 INFO HandlerThread:3594 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 18:58:47,866 DEBUG SenderThread:3594 [sender.py:send():378] send: final +2024-05-22 18:58:47,866 DEBUG SenderThread:3594 [sender.py:send():378] send: footer +2024-05-22 18:58:47,866 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: defer +2024-05-22 18:58:47,866 INFO SenderThread:3594 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 18:58:47,867 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:58:47,867 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:58:47,867 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 18:58:47,867 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 18:58:47,867 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 18:58:47,867 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 18:58:47,868 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:58:47,868 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:58:47,868 DEBUG SenderThread:3594 [sender.py:send_request():405] send_request: server_info +2024-05-22 18:58:47,920 INFO MainThread:3594 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 18:58:47,920 INFO MainThread:3594 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 18:58:47,920 INFO MainThread:3594 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-22 18:58:47,921 DEBUG HandlerThread:3594 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 18:58:47,921 INFO HandlerThread:3594 [handler.py:finish():882] shutting down handler +2024-05-22 18:58:48,868 INFO WriterThread:3594 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/run-auaa3wb6.wandb +2024-05-22 18:58:48,920 INFO SenderThread:3594 [sender.py:finish():1545] shutting down sender +2024-05-22 18:58:48,920 INFO SenderThread:3594 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:58:48,920 INFO SenderThread:3594 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..2b641839563442b415c5ea01de803e1d6291f5d8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 18:58:33,655 INFO MainThread:3439 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 18:58:33,655 INFO MainThread:3439 [wandb_setup.py:_flush():76] Configure stats pid to 3439 +2024-05-22 18:58:33,655 INFO MainThread:3439 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 18:58:33,655 INFO MainThread:3439 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 18:58:33,655 INFO MainThread:3439 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 18:58:33,655 INFO MainThread:3439 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 18:58:33,655 WARNING MainThread:3439 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 18:58:33,655 INFO MainThread:3439 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 18:58:33,656 INFO MainThread:3439 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 18:58:33,656 INFO MainThread:3439 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/logs/debug.log +2024-05-22 18:58:33,656 INFO MainThread:3439 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/logs/debug-internal.log +2024-05-22 18:58:33,656 INFO MainThread:3439 [wandb_init.py:init():560] calling init triggers +2024-05-22 18:58:33,656 INFO MainThread:3439 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 18:58:33,656 INFO MainThread:3439 [wandb_init.py:init():610] starting backend +2024-05-22 18:58:33,656 INFO MainThread:3439 [wandb_init.py:init():614] setting up manager +2024-05-22 18:58:33,660 INFO MainThread:3439 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 18:58:33,661 INFO MainThread:3439 [wandb_init.py:init():622] backend started and connected +2024-05-22 18:58:33,665 INFO MainThread:3439 [wandb_init.py:init():711] updated telemetry +2024-05-22 18:58:33,674 INFO MainThread:3439 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 18:58:33,979 INFO MainThread:3439 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 18:58:34,097 INFO MainThread:3439 [wandb_run.py:_on_init():2405] got version response +2024-05-22 18:58:34,098 INFO MainThread:3439 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 18:58:34,391 INFO MainThread:3439 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 18:58:34,391 INFO MainThread:3439 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 18:58:34,391 INFO MainThread:3439 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 18:58:34,391 INFO MainThread:3439 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 18:58:34,394 INFO MainThread:3439 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 18:58:48,922 WARNING MsgRouterThr:3439 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/run-auaa3wb6.wandb b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/run-auaa3wb6.wandb new file mode 100644 index 0000000000000000000000000000000000000000..c2f8e9c8e239bbb02c741160c42eba72e61864a2 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_185833-auaa3wb6/run-auaa3wb6.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..94aa8cc6a5b0c0a919fdc1ede42e4393f9a4e1e6 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716441284 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/output.log b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..51e091c32333207b5c80d280b38f7f9dd80fd520 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:05:14:44,686 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:05:14:49,187 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:05:14:49,188 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:05:14:49,188 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000'} +2024-05-23:05:14:50,424 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..7599f06c4054ceb59eeb416bac091b0e666613ef --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T05:14:44.557317", + "startedAt": "2024-05-23T05:14:44.090737", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2372.16768125, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 212.1846694946289 + } + }, + "memory": { + "total": 1007.43798828125 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..e682bae6b5eaeba8295fd0fffdc51474a259249e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 5}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..831de0efe4de37d366e821ddeb4e93a4e9de6c88 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/logs/debug-internal.log @@ -0,0 +1,179 @@ +2024-05-23 05:14:44,105 INFO StreamThr :8575 [internal.py:wandb_internal():85] W&B internal server running at pid: 8575, started at: 2024-05-23 05:14:44.104880 +2024-05-23 05:14:44,106 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: status +2024-05-23 05:14:44,107 INFO WriterThread:8575 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/run-rw690kmy.wandb +2024-05-23 05:14:44,108 DEBUG SenderThread:8575 [sender.py:send():378] send: header +2024-05-23 05:14:44,118 DEBUG SenderThread:8575 [sender.py:send():378] send: run +2024-05-23 05:14:44,410 INFO SenderThread:8575 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files +2024-05-23 05:14:44,410 INFO SenderThread:8575 [sender.py:_start_run_threads():1123] run started: rw690kmy with start time 1716441284.104686 +2024-05-23 05:14:44,416 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 05:14:44,416 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: check_version +2024-05-23 05:14:44,502 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 05:14:44,504 DEBUG HandlerThread:8575 [system_info.py:__init__():26] System info init +2024-05-23 05:14:44,504 DEBUG HandlerThread:8575 [system_info.py:__init__():41] System info init done +2024-05-23 05:14:44,504 INFO HandlerThread:8575 [system_monitor.py:start():194] Starting system monitor +2024-05-23 05:14:44,504 INFO SystemMonitor:8575 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 05:14:44,504 INFO HandlerThread:8575 [system_monitor.py:probe():214] Collecting system info +2024-05-23 05:14:44,505 INFO SystemMonitor:8575 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 05:14:44,505 INFO SystemMonitor:8575 [interfaces.py:start():188] Started disk monitoring +2024-05-23 05:14:44,506 INFO SystemMonitor:8575 [interfaces.py:start():188] Started memory monitoring +2024-05-23 05:14:44,507 INFO SystemMonitor:8575 [interfaces.py:start():188] Started network monitoring +2024-05-23 05:14:44,557 DEBUG HandlerThread:8575 [system_info.py:probe():150] Probing system +2024-05-23 05:14:44,567 DEBUG HandlerThread:8575 [system_info.py:_probe_git():135] Probing git +2024-05-23 05:14:44,585 ERROR HandlerThread:8575 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 05:14:44,585 DEBUG HandlerThread:8575 [system_info.py:_probe_git():143] Probing git done +2024-05-23 05:14:44,585 DEBUG HandlerThread:8575 [system_info.py:probe():198] Probing system done +2024-05-23 05:14:44,585 DEBUG HandlerThread:8575 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T05:14:44.557317', 'startedAt': '2024-05-23T05:14:44.090737', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step14000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2372.16768125, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 212.1846694946289}}, 'memory': {'total': 1007.43798828125}} +2024-05-23 05:14:44,585 INFO HandlerThread:8575 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 05:14:44,585 INFO HandlerThread:8575 [system_monitor.py:probe():227] Publishing system info +2024-05-23 05:14:44,587 INFO HandlerThread:8575 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 05:14:44,591 DEBUG SenderThread:8575 [sender.py:send():378] send: files +2024-05-23 05:14:44,591 INFO SenderThread:8575 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 05:14:44,682 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 05:14:44,682 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 05:14:44,682 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: python_packages +2024-05-23 05:14:44,685 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: stop_status +2024-05-23 05:14:44,857 DEBUG SenderThread:8575 [sender.py:send():378] send: telemetry +2024-05-23 05:14:45,126 INFO wandb-upload_0:8575 [upload_job.py:push():130] Uploaded file /tmp/tmp9gvhummywandb/20lft4ze-wandb-metadata.json +2024-05-23 05:14:45,411 INFO Thread-12 :8575 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/requirements.txt +2024-05-23 05:14:45,412 INFO Thread-12 :8575 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/output.log +2024-05-23 05:14:45,412 INFO Thread-12 :8575 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-metadata.json +2024-05-23 05:14:47,412 INFO Thread-12 :8575 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/output.log +2024-05-23 05:14:49,188 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:14:50,429 DEBUG SenderThread:8575 [sender.py:send():378] send: exit +2024-05-23 05:14:50,429 INFO SenderThread:8575 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 05:14:50,430 INFO SenderThread:8575 [sender.py:send_exit():587] handling runtime: 5 +2024-05-23 05:14:50,431 INFO SenderThread:8575 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:14:50,431 INFO SenderThread:8575 [sender.py:send_exit():593] send defer +2024-05-23 05:14:50,431 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,431 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 05:14:50,431 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,431 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 05:14:50,431 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 1 +2024-05-23 05:14:50,431 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,432 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 05:14:50,432 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,432 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 05:14:50,432 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 2 +2024-05-23 05:14:50,432 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,432 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 05:14:50,432 INFO HandlerThread:8575 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 05:14:50,433 DEBUG SystemMonitor:8575 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 05:14:50,433 DEBUG SystemMonitor:8575 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 05:14:50,433 DEBUG SystemMonitor:8575 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 05:14:50,433 INFO HandlerThread:8575 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 05:14:50,433 INFO HandlerThread:8575 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 05:14:50,433 INFO HandlerThread:8575 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 05:14:50,434 INFO HandlerThread:8575 [interfaces.py:finish():200] Joined network monitor +2024-05-23 05:14:50,434 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,434 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 05:14:50,434 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 3 +2024-05-23 05:14:50,434 DEBUG SenderThread:8575 [sender.py:send():378] send: stats +2024-05-23 05:14:50,434 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,434 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 05:14:50,434 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,434 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 05:14:50,434 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 4 +2024-05-23 05:14:50,434 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,434 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 05:14:50,434 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,434 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 05:14:50,434 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 5 +2024-05-23 05:14:50,435 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,435 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 05:14:50,435 DEBUG SenderThread:8575 [sender.py:send():378] send: summary +2024-05-23 05:14:50,435 INFO SenderThread:8575 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:14:50,436 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,436 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 05:14:50,436 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 6 +2024-05-23 05:14:50,436 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,436 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 05:14:50,436 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,436 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 05:14:50,438 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:14:50,525 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 7 +2024-05-23 05:14:50,525 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,525 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 05:14:50,525 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,525 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 05:14:50,876 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 8 +2024-05-23 05:14:50,876 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,876 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 05:14:50,876 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,876 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 05:14:50,876 INFO SenderThread:8575 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 05:14:50,877 INFO SenderThread:8575 [job_builder.py:_get_source_type():576] no source found +2024-05-23 05:14:50,877 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 9 +2024-05-23 05:14:50,877 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:50,877 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 05:14:50,877 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:50,877 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 05:14:50,877 INFO SenderThread:8575 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 05:14:51,414 INFO SenderThread:8575 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/config.yaml +2024-05-23 05:14:51,414 INFO SenderThread:8575 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/output.log +2024-05-23 05:14:51,414 INFO SenderThread:8575 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-summary.json +2024-05-23 05:14:51,414 INFO SenderThread:8575 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files +2024-05-23 05:14:51,414 INFO SenderThread:8575 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/requirements.txt requirements.txt +2024-05-23 05:14:51,414 INFO SenderThread:8575 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-summary.json wandb-summary.json +2024-05-23 05:14:51,416 INFO SenderThread:8575 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/output.log output.log +2024-05-23 05:14:51,417 INFO SenderThread:8575 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-metadata.json wandb-metadata.json +2024-05-23 05:14:51,417 INFO SenderThread:8575 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/config.yaml config.yaml +2024-05-23 05:14:51,418 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 10 +2024-05-23 05:14:51,419 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:51,419 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 05:14:51,420 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:51,420 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 05:14:51,420 INFO SenderThread:8575 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:14:51,429 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:14:51,430 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:14:51,653 INFO wandb-upload_0:8575 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/requirements.txt +2024-05-23 05:14:51,837 INFO wandb-upload_1:8575 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/wandb-summary.json +2024-05-23 05:14:51,920 INFO wandb-upload_2:8575 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/output.log +2024-05-23 05:14:51,933 INFO wandb-upload_3:8575 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/files/config.yaml +2024-05-23 05:14:52,133 INFO Thread-11 (_thread_body):8575 [sender.py:transition_state():613] send defer: 11 +2024-05-23 05:14:52,133 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:52,133 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 05:14:52,134 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:52,134 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 05:14:52,134 INFO SenderThread:8575 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 05:14:52,134 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 12 +2024-05-23 05:14:52,134 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:52,134 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 05:14:52,134 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:52,134 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 05:14:52,134 INFO SenderThread:8575 [file_stream.py:finish():601] file stream finish called +2024-05-23 05:14:52,349 INFO SenderThread:8575 [file_stream.py:finish():605] file stream finish is done +2024-05-23 05:14:52,349 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 13 +2024-05-23 05:14:52,349 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:52,349 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 05:14:52,349 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:52,349 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 05:14:52,349 INFO SenderThread:8575 [sender.py:transition_state():613] send defer: 14 +2024-05-23 05:14:52,349 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:14:52,350 INFO HandlerThread:8575 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 05:14:52,350 DEBUG SenderThread:8575 [sender.py:send():378] send: final +2024-05-23 05:14:52,350 DEBUG SenderThread:8575 [sender.py:send():378] send: footer +2024-05-23 05:14:52,350 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: defer +2024-05-23 05:14:52,350 INFO SenderThread:8575 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 05:14:52,350 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:14:52,350 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:14:52,351 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:14:52,351 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:14:52,351 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 05:14:52,351 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 05:14:52,351 DEBUG SenderThread:8575 [sender.py:send_request():405] send_request: server_info +2024-05-23 05:14:52,353 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 05:14:52,353 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 05:14:52,414 INFO MainThread:8575 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 05:14:52,414 INFO MainThread:8575 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 05:14:52,414 INFO MainThread:8575 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 05:14:52,414 DEBUG HandlerThread:8575 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 05:14:52,415 INFO HandlerThread:8575 [handler.py:finish():882] shutting down handler +2024-05-23 05:14:53,351 INFO WriterThread:8575 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/run-rw690kmy.wandb +2024-05-23 05:14:53,414 INFO SenderThread:8575 [sender.py:finish():1545] shutting down sender +2024-05-23 05:14:53,414 INFO SenderThread:8575 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:14:53,414 INFO SenderThread:8575 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..120111ead82421fe28ed158a5b736ccc8a370f18 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 05:14:44,101 INFO MainThread:8420 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 05:14:44,101 INFO MainThread:8420 [wandb_setup.py:_flush():76] Configure stats pid to 8420 +2024-05-23 05:14:44,101 INFO MainThread:8420 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 05:14:44,101 INFO MainThread:8420 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 05:14:44,102 WARNING MainThread:8420 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/logs/debug.log +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/logs/debug-internal.log +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_init.py:init():560] calling init triggers +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_init.py:init():610] starting backend +2024-05-23 05:14:44,102 INFO MainThread:8420 [wandb_init.py:init():614] setting up manager +2024-05-23 05:14:44,103 INFO MainThread:8420 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 05:14:44,104 INFO MainThread:8420 [wandb_init.py:init():622] backend started and connected +2024-05-23 05:14:44,106 INFO MainThread:8420 [wandb_init.py:init():711] updated telemetry +2024-05-23 05:14:44,117 INFO MainThread:8420 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 05:14:44,415 INFO MainThread:8420 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 05:14:44,498 INFO MainThread:8420 [wandb_run.py:_on_init():2405] got version response +2024-05-23 05:14:44,498 INFO MainThread:8420 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 05:14:44,683 INFO MainThread:8420 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 05:14:44,683 INFO MainThread:8420 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 05:14:44,683 INFO MainThread:8420 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 05:14:44,683 INFO MainThread:8420 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 05:14:44,684 INFO MainThread:8420 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 05:14:53,415 WARNING MsgRouterThr:8420 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/run-rw690kmy.wandb b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/run-rw690kmy.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e5164ebef73cf0199af5736a472ad6230449e780 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_051444-rw690kmy/run-rw690kmy.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6f0ae5f48c48e9747ab63a5aac971b192be445f2 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716451343 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..fddb0b8cf0fb7c7ecb0312f9d0eb86592cdf8eaf --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:08:02:24,596 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:08:02:33,036 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:08:02:33,037 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:08:02:33,037 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step26000'} +2024-05-23:08:02:35,334 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step26000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step26000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..d393bd4dff8d381e9544160d24c91cb1277bd381 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T08:02:24.392553", + "startedAt": "2024-05-23T08:02:23.876943", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step26000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2333.5540812500003, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3394.858, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64032745361328 + } + }, + "memory": { + "total": 1007.4379806518555 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..7fd81be4b402de29e99e08f9e1abdbc0ed79b124 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 08:02:23,899 INFO StreamThr :3319 [internal.py:wandb_internal():85] W&B internal server running at pid: 3319, started at: 2024-05-23 08:02:23.896822 +2024-05-23 08:02:23,903 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: status +2024-05-23 08:02:23,904 INFO WriterThread:3319 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/run-jxqzpk69.wandb +2024-05-23 08:02:23,906 DEBUG SenderThread:3319 [sender.py:send():378] send: header +2024-05-23 08:02:23,909 DEBUG SenderThread:3319 [sender.py:send():378] send: run +2024-05-23 08:02:24,205 INFO SenderThread:3319 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files +2024-05-23 08:02:24,205 INFO SenderThread:3319 [sender.py:_start_run_threads():1123] run started: jxqzpk69 with start time 1716451343.896997 +2024-05-23 08:02:24,206 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 08:02:24,206 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: check_version +2024-05-23 08:02:24,325 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 08:02:24,327 DEBUG HandlerThread:3319 [system_info.py:__init__():26] System info init +2024-05-23 08:02:24,327 DEBUG HandlerThread:3319 [system_info.py:__init__():41] System info init done +2024-05-23 08:02:24,327 INFO HandlerThread:3319 [system_monitor.py:start():194] Starting system monitor +2024-05-23 08:02:24,327 INFO SystemMonitor:3319 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 08:02:24,327 INFO HandlerThread:3319 [system_monitor.py:probe():214] Collecting system info +2024-05-23 08:02:24,357 INFO SystemMonitor:3319 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 08:02:24,357 INFO SystemMonitor:3319 [interfaces.py:start():188] Started disk monitoring +2024-05-23 08:02:24,357 INFO SystemMonitor:3319 [interfaces.py:start():188] Started memory monitoring +2024-05-23 08:02:24,358 INFO SystemMonitor:3319 [interfaces.py:start():188] Started network monitoring +2024-05-23 08:02:24,392 DEBUG HandlerThread:3319 [system_info.py:probe():150] Probing system +2024-05-23 08:02:24,395 DEBUG HandlerThread:3319 [system_info.py:_probe_git():135] Probing git +2024-05-23 08:02:24,405 ERROR HandlerThread:3319 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 08:02:24,405 DEBUG HandlerThread:3319 [system_info.py:_probe_git():143] Probing git done +2024-05-23 08:02:24,405 DEBUG HandlerThread:3319 [system_info.py:probe():198] Probing system done +2024-05-23 08:02:24,405 DEBUG HandlerThread:3319 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T08:02:24.392553', 'startedAt': '2024-05-23T08:02:23.876943', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step26000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2333.5540812500003, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3394.858, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64032745361328}}, 'memory': {'total': 1007.4379806518555}} +2024-05-23 08:02:24,405 INFO HandlerThread:3319 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 08:02:24,405 INFO HandlerThread:3319 [system_monitor.py:probe():227] Publishing system info +2024-05-23 08:02:24,408 INFO HandlerThread:3319 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 08:02:24,413 DEBUG SenderThread:3319 [sender.py:send():378] send: files +2024-05-23 08:02:24,413 INFO SenderThread:3319 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 08:02:24,590 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 08:02:24,591 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: python_packages +2024-05-23 08:02:24,591 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 08:02:24,597 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: stop_status +2024-05-23 08:02:24,762 DEBUG SenderThread:3319 [sender.py:send():378] send: telemetry +2024-05-23 08:02:25,112 INFO wandb-upload_0:3319 [upload_job.py:push():130] Uploaded file /tmp/tmpbevo3cjkwandb/wahas1j6-wandb-metadata.json +2024-05-23 08:02:25,206 INFO Thread-12 :3319 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/requirements.txt +2024-05-23 08:02:25,207 INFO Thread-12 :3319 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log +2024-05-23 08:02:25,207 INFO Thread-12 :3319 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-metadata.json +2024-05-23 08:02:27,206 INFO Thread-12 :3319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log +2024-05-23 08:02:29,774 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:02:35,038 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:02:35,213 INFO Thread-12 :3319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log +2024-05-23 08:02:35,345 DEBUG SenderThread:3319 [sender.py:send():378] send: exit +2024-05-23 08:02:35,345 INFO SenderThread:3319 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 08:02:35,346 INFO SenderThread:3319 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 08:02:35,348 INFO SenderThread:3319 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 08:02:35,348 INFO SenderThread:3319 [sender.py:send_exit():593] send defer +2024-05-23 08:02:35,348 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:35,348 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 08:02:35,348 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:35,348 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 08:02:35,348 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 1 +2024-05-23 08:02:35,348 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:35,348 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 08:02:35,348 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:35,349 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 08:02:35,349 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 2 +2024-05-23 08:02:35,349 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:35,349 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 08:02:35,349 INFO HandlerThread:3319 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 08:02:35,349 DEBUG SystemMonitor:3319 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 08:02:35,349 DEBUG SystemMonitor:3319 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 08:02:35,349 DEBUG SystemMonitor:3319 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 08:02:35,350 INFO HandlerThread:3319 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 08:02:35,350 INFO HandlerThread:3319 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 08:02:35,350 INFO HandlerThread:3319 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 08:02:35,350 INFO HandlerThread:3319 [interfaces.py:finish():200] Joined network monitor +2024-05-23 08:02:35,351 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:35,351 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 08:02:35,351 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 3 +2024-05-23 08:02:35,351 DEBUG SenderThread:3319 [sender.py:send():378] send: stats +2024-05-23 08:02:35,351 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:35,351 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 08:02:35,351 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:35,351 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 08:02:35,351 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 4 +2024-05-23 08:02:35,351 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:35,351 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 08:02:35,351 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:35,351 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 08:02:35,351 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 5 +2024-05-23 08:02:35,352 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:35,352 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 08:02:35,352 DEBUG SenderThread:3319 [sender.py:send():378] send: summary +2024-05-23 08:02:35,352 INFO SenderThread:3319 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 08:02:35,353 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:35,353 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 08:02:35,353 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 6 +2024-05-23 08:02:35,353 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:35,353 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 08:02:35,353 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:35,353 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 08:02:35,357 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 08:02:35,441 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 7 +2024-05-23 08:02:35,441 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:35,441 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 08:02:35,441 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:35,441 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 08:02:36,214 INFO Thread-12 :3319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/config.yaml +2024-05-23 08:02:36,215 INFO Thread-12 :3319 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-summary.json +2024-05-23 08:02:36,346 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:02:36,793 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 8 +2024-05-23 08:02:36,793 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:02:36,793 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:36,793 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 08:02:36,794 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:36,794 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 08:02:36,794 INFO SenderThread:3319 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 08:02:36,794 INFO SenderThread:3319 [job_builder.py:_get_source_type():576] no source found +2024-05-23 08:02:36,794 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 9 +2024-05-23 08:02:36,794 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:36,794 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 08:02:36,794 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:36,795 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 08:02:36,795 INFO SenderThread:3319 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 08:02:37,216 INFO SenderThread:3319 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log +2024-05-23 08:02:37,216 INFO SenderThread:3319 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files +2024-05-23 08:02:37,216 INFO SenderThread:3319 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-summary.json wandb-summary.json +2024-05-23 08:02:37,216 INFO SenderThread:3319 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/requirements.txt requirements.txt +2024-05-23 08:02:37,218 INFO SenderThread:3319 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-metadata.json wandb-metadata.json +2024-05-23 08:02:37,219 INFO SenderThread:3319 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log output.log +2024-05-23 08:02:37,219 INFO SenderThread:3319 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/config.yaml config.yaml +2024-05-23 08:02:37,219 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 10 +2024-05-23 08:02:37,219 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:37,219 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 08:02:37,219 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:37,219 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 08:02:37,219 INFO SenderThread:3319 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 08:02:37,346 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:02:37,346 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:02:37,543 INFO wandb-upload_0:3319 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/wandb-summary.json +2024-05-23 08:02:37,808 INFO wandb-upload_2:3319 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/output.log +2024-05-23 08:02:37,825 INFO wandb-upload_1:3319 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/requirements.txt +2024-05-23 08:02:37,836 INFO wandb-upload_3:3319 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/files/config.yaml +2024-05-23 08:02:38,036 INFO Thread-11 (_thread_body):3319 [sender.py:transition_state():613] send defer: 11 +2024-05-23 08:02:38,037 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:38,037 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 08:02:38,037 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:38,037 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 08:02:38,037 INFO SenderThread:3319 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 08:02:38,037 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 12 +2024-05-23 08:02:38,037 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:38,037 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 08:02:38,037 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:38,037 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 08:02:38,037 INFO SenderThread:3319 [file_stream.py:finish():601] file stream finish called +2024-05-23 08:02:38,111 INFO SenderThread:3319 [file_stream.py:finish():605] file stream finish is done +2024-05-23 08:02:38,111 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 13 +2024-05-23 08:02:38,111 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:38,111 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 08:02:38,112 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:38,112 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 08:02:38,112 INFO SenderThread:3319 [sender.py:transition_state():613] send defer: 14 +2024-05-23 08:02:38,112 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: defer +2024-05-23 08:02:38,112 INFO HandlerThread:3319 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 08:02:38,112 DEBUG SenderThread:3319 [sender.py:send():378] send: final +2024-05-23 08:02:38,112 DEBUG SenderThread:3319 [sender.py:send():378] send: footer +2024-05-23 08:02:38,112 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: defer +2024-05-23 08:02:38,112 INFO SenderThread:3319 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 08:02:38,113 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:02:38,113 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 08:02:38,113 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 08:02:38,113 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 08:02:38,113 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 08:02:38,113 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 08:02:38,113 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:02:38,114 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 08:02:38,114 DEBUG SenderThread:3319 [sender.py:send_request():405] send_request: server_info +2024-05-23 08:02:38,166 INFO MainThread:3319 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 08:02:38,166 INFO MainThread:3319 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 08:02:38,167 INFO MainThread:3319 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 08:02:38,167 DEBUG HandlerThread:3319 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 08:02:38,167 INFO HandlerThread:3319 [handler.py:finish():882] shutting down handler +2024-05-23 08:02:39,114 INFO WriterThread:3319 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/run-jxqzpk69.wandb +2024-05-23 08:02:39,166 INFO SenderThread:3319 [sender.py:finish():1545] shutting down sender +2024-05-23 08:02:39,166 INFO SenderThread:3319 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 08:02:39,167 INFO SenderThread:3319 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..e4ac6babf17eaf16f9d594bc58ec3a4d9154a1e3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_setup.py:_flush():76] Configure stats pid to 3164 +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 08:02:23,891 WARNING MainThread:3164 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/logs/debug.log +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/logs/debug-internal.log +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_init.py:init():560] calling init triggers +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_init.py:init():610] starting backend +2024-05-23 08:02:23,891 INFO MainThread:3164 [wandb_init.py:init():614] setting up manager +2024-05-23 08:02:23,895 INFO MainThread:3164 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 08:02:23,896 INFO MainThread:3164 [wandb_init.py:init():622] backend started and connected +2024-05-23 08:02:23,900 INFO MainThread:3164 [wandb_init.py:init():711] updated telemetry +2024-05-23 08:02:23,908 INFO MainThread:3164 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 08:02:24,206 INFO MainThread:3164 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 08:02:24,318 INFO MainThread:3164 [wandb_run.py:_on_init():2405] got version response +2024-05-23 08:02:24,318 INFO MainThread:3164 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 08:02:24,592 INFO MainThread:3164 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 08:02:24,592 INFO MainThread:3164 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 08:02:24,592 INFO MainThread:3164 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 08:02:24,592 INFO MainThread:3164 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 08:02:24,594 INFO MainThread:3164 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 08:02:39,168 WARNING MsgRouterThr:3164 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/run-jxqzpk69.wandb b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/run-jxqzpk69.wandb new file mode 100644 index 0000000000000000000000000000000000000000..da3a95252934564bcba072b2cdd952e843093bb3 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_080223-jxqzpk69/run-jxqzpk69.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/config.yaml b/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..6f0d46de3f78e4b388b6ac52cc9bfaec2e263f57 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/config.yaml @@ -0,0 +1,283 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.36.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716987506 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 2 + - 23 + - 62 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.36.2 + 8: + - 5 + 13: linux-x86_64 +task_configs: + desc: null + value: + arc_easy: + task: arc_easy + group: + - ai2_arc + dataset_path: allenai/ai2_arc + dataset_name: ARC-Easy + training_split: train + validation_split: validation + test_split: test + doc_to_text: 'Question: {{question}} + + Answer:' + doc_to_target: '{{choices.label.index(answerKey)}}' + doc_to_choice: '{{choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{question}} + + Answer:' + metadata: + version: 1.0 + boolq: + task: boolq + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: boolq + training_split: train + validation_split: validation + doc_to_text: '{{passage}} + + Question: {{question}}? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: passage + metadata: + version: 2.0 + copa: + task: copa + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: copa + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n # Drop the period\n connector =\ + \ {\n \"cause\": \"because\",\n \"effect\": \"therefore\",\n\ + \ }[doc[\"question\"]]\n return doc[\"premise\"].strip()[:-1] + f\"\ + \ {connector}\"\n" + doc_to_target: "def doc_to_target(doc):\n correct_choice = doc[\"choice1\"\ + ] if doc[\"label\"] == 0 else doc[\"choice2\"]\n # Connect the sentences\n\ + \ return \" \" + convert_choice(correct_choice)\n" + doc_to_choice: "def doc_to_choice(doc):\n return [\" \" + convert_choice(doc[\"\ + choice1\"]), \" \" + convert_choice(doc[\"choice2\"])]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + mrpc: + task: mrpc + group: glue + dataset_path: glue + dataset_name: mrpc + training_split: train + validation_split: validation + doc_to_text: 'Sentence 1: {{sentence1}} + + Sentence 2: {{sentence2}} + + Question: Do both sentences mean the same thing? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + - metric: f1 + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + piqa: + task: piqa + dataset_path: piqa + training_split: train + validation_split: validation + doc_to_text: 'Question: {{goal}} + + Answer:' + doc_to_target: label + doc_to_choice: '{{[sol1, sol2]}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: goal + metadata: + version: 1.0 + sst2: + task: sst2 + group: glue + dataset_path: glue + dataset_name: sst2 + training_split: train + validation_split: validation + doc_to_text: '{{sentence}} + + Question: Is this sentence positive or negative? + + Answer:' + doc_to_target: label + doc_to_choice: + - negative + - positive + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + winogrande: + task: winogrande + dataset_path: winogrande + dataset_name: winogrande_xl + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n answer_to_num = {\"1\": 0, \"2\": 1}\n\ + \ return answer_to_num[doc[\"answer\"]]\n" + doc_to_target: "def doc_to_target(doc):\n idx = doc[\"sentence\"].index(\"\ + _\") + 1\n return doc[\"sentence\"][idx:].strip()\n" + doc_to_choice: "def doc_to_choice(doc):\n idx = doc[\"sentence\"].index(\"\ + _\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"\ + sentence\"][:idx] + opt for opt in options]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: sentence + metadata: + version: 1.0 +cli_configs: + desc: null + value: + model: hf + model_args: pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint/llamav2-3b/hf/global_step10000,tokenizer=huggyllama/llama-13b + batch_size: auto + batch_sizes: + - 64 + device: null + use_cache: null + limit: null + bootstrap_iters: 100000 + gen_kwargs: null diff --git a/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/media/table/evaluation/eval_results_1_e9c0c28833db92418bd4.table.json b/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/media/table/evaluation/eval_results_1_e9c0c28833db92418bd4.table.json new file mode 100644 index 0000000000000000000000000000000000000000..c81f60d564d216dfe579f8ecfdf11573cd71302e --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/media/table/evaluation/eval_results_1_e9c0c28833db92418bd4.table.json @@ -0,0 +1 @@ +{"columns": ["Tasks", "Version", "Filter", "num_fewshot", "Metric", "Value", "Stderr"], "data": [["winogrande", 1.0, "none", 0, "acc", "0.49013417521704816", "0.0140"], ["sst2", 1.0, "none", 0, "acc", "0.4954128440366973", "0.0169"], ["piqa", 1.0, "none", 0, "acc", "0.5223068552774756", "0.0117"], ["piqa", 1.0, "none", 0, "acc_norm", "0.49891186071817195", "0.0117"], ["mrpc", 1.0, "none", 0, "acc", "0.5955882352941176", "0.0243"], ["mrpc", 1.0, "none", 0, "f1", "0.7140381282495667", "0.0214"], ["copa", 1.0, "none", 0, "acc", "0.63", "0.0485"], ["boolq", 2.0, "none", 0, "acc", "0.5214067278287462", "0.0087"], ["arc_easy", 1.0, "none", 0, "acc", "0.24621212121212122", "0.0088"], ["arc_easy", 1.0, "none", 0, "acc_norm", "0.2521043771043771", "0.0089"]]} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/output.log b/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..5b6d01369fdd58bb02b3901db10e3e484c89dd56 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240529_125826-ol4022ex/files/output.log @@ -0,0 +1,477 @@ + +2024-05-29:12:58:26,910 INFO [__main__.py:251] Verbosity set to INFO +2024-05-29:12:58:36,393 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'boolq', 'copa', 'mrpc', 'piqa', 'sst2', 'winogrande'] +2024-05-29:12:58:36,395 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-29:12:58:36,395 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/eval/checkpoint/llamav2-3b/hf/global_step10000', 'tokenizer': 'huggyllama/llama-13b'} +2024-05-29:12:58:38,963 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +/usr/local/lib/python3.10/dist-packages/huggingface_hub/file_download.py:1132: FutureWarning: `resume_download` is deprecated and will be removed in version 1.0.0. Downloads always resume when possible. If you want to force a new download, use `force_download=True`. + warnings.warn( +Downloading readme: 100%|██████████| 9.00k/9.00k [00:00<00:00, 16.9MB/s] +Downloading data: 100%|██████████| 331k/331k [00:00<00:00, 450kB/s] +Downloading data: 100%|██████████| 346k/346k [00:00<00:00, 2.94MB/s] +Downloading data: 100%|██████████| 86.1k/86.1k [00:00<00:00, 728kB/s] +Generating train split: 100%|██████████| 2251/2251 [00:00<00:00, 56848.03 examples/s] +Generating test split: 100%|██████████| 2376/2376 [00:00<00:00, 326208.39 examples/s] +Generating validation split: 100%|██████████| 570/570 [00:00<00:00, 191046.29 examples/s] +2024-05-29:12:59:10,613 WARNING [task.py:763] [Task: boolq] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-29:12:59:10,614 WARNING [task.py:775] [Task: boolq] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for super_glue contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/super_glue +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 30.7k/30.7k [00:00<00:00, 42.2MB/s] +Downloading readme: 100%|██████████| 18.2k/18.2k [00:00<00:00, 30.6MB/s] +Downloading data: 100%|██████████| 4.12M/4.12M [00:00<00:00, 82.8MB/s] +Generating train split: 100%|██████████| 9427/9427 [00:00<00:00, 22440.78 examples/s] +Generating validation split: 100%|██████████| 3270/3270 [00:00<00:00, 22663.93 examples/s] +Generating test split: 100%|██████████| 3245/3245 [00:00<00:00, 23659.31 examples/s] +2024-05-29:12:59:14,743 WARNING [task.py:763] [Task: copa] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-29:12:59:14,744 WARNING [task.py:775] [Task: copa] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +Downloading data: 100%|██████████| 44.0k/44.0k [00:00<00:00, 47.6MB/s] +Generating train split: 100%|██████████| 400/400 [00:00<00:00, 16168.32 examples/s] +Generating validation split: 100%|██████████| 100/100 [00:00<00:00, 12773.88 examples/s] +Generating test split: 100%|██████████| 500/500 [00:00<00:00, 17231.72 examples/s] +2024-05-29:12:59:16,826 WARNING [task.py:763] [Task: mrpc] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-29:12:59:16,826 WARNING [task.py:775] [Task: mrpc] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +2024-05-29:12:59:16,827 WARNING [task.py:763] [Task: mrpc] metric f1 is defined, but aggregation is not. using default aggregation=f1 +2024-05-29:12:59:16,827 WARNING [task.py:775] [Task: mrpc] metric f1 is defined, but higher_is_better is not. using default higher_is_better=True +Downloading readme: 100%|██████████| 35.3k/35.3k [00:00<00:00, 646kB/s] +Downloading data: 100%|██████████| 649k/649k [00:00<00:00, 3.90MB/s] +Downloading data: 100%|██████████| 75.7k/75.7k [00:00<00:00, 501kB/s] +Downloading data: 100%|██████████| 308k/308k [00:00<00:00, 2.10MB/s] +Generating train split: 100%|██████████| 3668/3668 [00:00<00:00, 401354.14 examples/s] +Generating validation split: 100%|██████████| 408/408 [00:00<00:00, 175677.65 examples/s] +Generating test split: 100%|██████████| 1725/1725 [00:00<00:00, 361379.27 examples/s] +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for piqa contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/piqa +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 5.36k/5.36k [00:00<00:00, 11.8MB/s] +Downloading readme: 100%|██████████| 8.41k/8.41k [00:00<00:00, 17.6MB/s] +Downloading data: 100%|██████████| 1.82M/1.82M [00:00<00:00, 4.31MB/s] +Downloading data: 100%|██████████| 815k/815k [00:00<00:00, 23.0MB/s] +Generating train split: 100%|██████████| 16113/16113 [00:00<00:00, 24004.05 examples/s] +Generating test split: 100%|██████████| 3084/3084 [00:00<00:00, 24311.61 examples/s] +Generating validation split: 100%|██████████| 1838/1838 [00:00<00:00, 23948.17 examples/s] +2024-05-29:12:59:27,775 WARNING [task.py:763] [Task: sst2] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-05-29:12:59:27,776 WARNING [task.py:775] [Task: sst2] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +Downloading data: 100%|██████████| 3.11M/3.11M [00:00<00:00, 15.7MB/s] +Downloading data: 100%|██████████| 72.8k/72.8k [00:00<00:00, 269kB/s] +Downloading data: 100%|██████████| 148k/148k [00:00<00:00, 1.01MB/s] +Generating train split: 100%|██████████| 67349/67349 [00:00<00:00, 1417186.82 examples/s] +Generating validation split: 100%|██████████| 872/872 [00:00<00:00, 376899.54 examples/s] +Generating test split: 100%|██████████| 1821/1821 [00:00<00:00, 550394.72 examples/s] +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1486: FutureWarning: The repository for winogrande contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/winogrande +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +Downloading builder script: 100%|██████████| 5.65k/5.65k [00:00<00:00, 12.4MB/s] +Downloading readme: 100%|██████████| 9.97k/9.97k [00:00<00:00, 20.8MB/s] +Downloading data: 100%|██████████| 3.40M/3.40M [00:00<00:00, 7.06MB/s] +Generating train split: 100%|██████████| 40398/40398 [00:01<00:00, 24569.19 examples/s] +Generating test split: 100%|██████████| 1767/1767 [00:00<00:00, 24571.76 examples/s] +Generating validation split: 100%|██████████| 1267/1267 [00:00<00:00, 24004.04 examples/s] +2024-05-29:12:59:41,227 INFO [task.py:395] Building contexts for winogrande on rank 0... +100%|██████████| 1267/1267 [00:00<00:00, 68346.99it/s] +2024-05-29:12:59:41,312 INFO [task.py:395] Building contexts for sst2 on rank 0... +100%|██████████| 872/872 [00:00<00:00, 2532.37it/s] +2024-05-29:12:59:41,685 INFO [task.py:395] Building contexts for piqa on rank 0... +100%|██████████| 1838/1838 [00:01<00:00, 1069.87it/s] +2024-05-29:12:59:43,496 INFO [task.py:395] Building contexts for mrpc on rank 0... +100%|██████████| 408/408 [00:00<00:00, 1876.85it/s] +2024-05-29:12:59:43,746 INFO [task.py:395] Building contexts for copa on rank 0... +100%|██████████| 100/100 [00:00<00:00, 60857.57it/s] +2024-05-29:12:59:43,755 INFO [task.py:395] Building contexts for boolq on rank 0... +100%|██████████| 3270/3270 [00:01<00:00, 1997.40it/s] +2024-05-29:12:59:45,532 INFO [task.py:395] Building contexts for arc_easy on rank 0... +100%|██████████| 2376/2376 [00:02<00:00, 1063.28it/s] +2024-05-29:12:59:47,914 INFO [evaluator.py:379] Running loglikelihood requests +Running loglikelihood requests: 0%| | 0/25011 [00:00 1024). Running this sequence through the model will result in indexing errors +Running loglikelihood requests: 0%| | 0/45739 [00:00