diff --git a/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..27e7d41dbeeaadf4ba4cb83c584c0430f15a5ce3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/logs/debug-internal.log @@ -0,0 +1,182 @@ +2024-05-14 16:36:06,857 INFO StreamThr :125128 [internal.py:wandb_internal():85] W&B internal server running at pid: 125128, started at: 2024-05-14 16:36:06.856890 +2024-05-14 16:36:06,859 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: status +2024-05-14 16:36:06,861 INFO WriterThread:125128 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/run-tnpwn56x.wandb +2024-05-14 16:36:06,861 DEBUG SenderThread:125128 [sender.py:send():378] send: header +2024-05-14 16:36:06,870 DEBUG SenderThread:125128 [sender.py:send():378] send: run +2024-05-14 16:36:07,099 INFO SenderThread:125128 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files +2024-05-14 16:36:07,100 INFO SenderThread:125128 [sender.py:_start_run_threads():1123] run started: tnpwn56x with start time 1715704566.856692 +2024-05-14 16:36:07,106 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 16:36:07,107 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: check_version +2024-05-14 16:36:07,188 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 16:36:07,190 DEBUG HandlerThread:125128 [system_info.py:__init__():26] System info init +2024-05-14 16:36:07,190 DEBUG HandlerThread:125128 [system_info.py:__init__():41] System info init done +2024-05-14 16:36:07,190 INFO HandlerThread:125128 [system_monitor.py:start():194] Starting system monitor +2024-05-14 16:36:07,190 INFO SystemMonitor:125128 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 16:36:07,190 INFO HandlerThread:125128 [system_monitor.py:probe():214] Collecting system info +2024-05-14 16:36:07,191 INFO SystemMonitor:125128 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 16:36:07,192 INFO SystemMonitor:125128 [interfaces.py:start():188] Started disk monitoring +2024-05-14 16:36:07,192 INFO SystemMonitor:125128 [interfaces.py:start():188] Started memory monitoring +2024-05-14 16:36:07,192 INFO SystemMonitor:125128 [interfaces.py:start():188] Started network monitoring +2024-05-14 16:36:07,256 DEBUG HandlerThread:125128 [system_info.py:probe():150] Probing system +2024-05-14 16:36:07,265 DEBUG HandlerThread:125128 [system_info.py:_probe_git():135] Probing git +2024-05-14 16:36:07,285 ERROR HandlerThread:125128 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 16:36:07,285 DEBUG HandlerThread:125128 [system_info.py:_probe_git():143] Probing git done +2024-05-14 16:36:07,285 DEBUG HandlerThread:125128 [system_info.py:probe():198] Probing system done +2024-05-14 16:36:07,285 DEBUG HandlerThread:125128 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T16:36:07.256791', 'startedAt': '2024-05-14T16:36:06.846253', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3390.5422302631573, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3221.194, 'min': 800.0, 'max': 3400.0}, {'current': 3299.942, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3277.874, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3297.999, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3228.113, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3280.784, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 863.4254150390625}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 16:36:07,285 INFO HandlerThread:125128 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 16:36:07,285 INFO HandlerThread:125128 [system_monitor.py:probe():227] Publishing system info +2024-05-14 16:36:07,287 INFO HandlerThread:125128 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 16:36:07,293 DEBUG SenderThread:125128 [sender.py:send():378] send: files +2024-05-14 16:36:07,293 INFO SenderThread:125128 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 16:36:07,388 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 16:36:07,388 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:36:07,388 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: python_packages +2024-05-14 16:36:07,393 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:36:07,591 DEBUG SenderThread:125128 [sender.py:send():378] send: telemetry +2024-05-14 16:36:07,811 INFO wandb-upload_0:125128 [upload_job.py:push():130] Uploaded file /tmp/tmpasy_rhrtwandb/4jnsv4yr-wandb-metadata.json +2024-05-14 16:36:08,101 INFO Thread-12 :125128 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/output.log +2024-05-14 16:36:08,101 INFO Thread-12 :125128 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/wandb-metadata.json +2024-05-14 16:36:08,101 INFO Thread-12 :125128 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/requirements.txt +2024-05-14 16:36:10,101 INFO Thread-12 :125128 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/output.log +2024-05-14 16:36:12,143 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:36:13,230 DEBUG SenderThread:125128 [sender.py:send():378] send: exit +2024-05-14 16:36:13,230 INFO SenderThread:125128 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 16:36:13,230 INFO SenderThread:125128 [sender.py:send_exit():587] handling runtime: 6 +2024-05-14 16:36:13,231 INFO SenderThread:125128 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:36:13,231 INFO SenderThread:125128 [sender.py:send_exit():593] send defer +2024-05-14 16:36:13,231 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:13,231 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 16:36:13,231 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:13,231 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 16:36:13,231 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 1 +2024-05-14 16:36:13,231 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:13,231 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 16:36:13,231 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:13,231 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 16:36:13,231 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 2 +2024-05-14 16:36:13,231 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:13,231 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 16:36:13,231 INFO HandlerThread:125128 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 16:36:13,232 INFO HandlerThread:125128 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 16:36:13,232 DEBUG SystemMonitor:125128 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 16:36:13,232 INFO HandlerThread:125128 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 16:36:13,232 DEBUG SystemMonitor:125128 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 16:36:13,232 INFO HandlerThread:125128 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 16:36:13,232 DEBUG SystemMonitor:125128 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 16:36:13,232 INFO HandlerThread:125128 [interfaces.py:finish():200] Joined network monitor +2024-05-14 16:36:13,234 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:13,234 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 16:36:13,234 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 3 +2024-05-14 16:36:13,234 DEBUG SenderThread:125128 [sender.py:send():378] send: stats +2024-05-14 16:36:13,234 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:13,235 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 16:36:13,235 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:13,235 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 16:36:13,235 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 4 +2024-05-14 16:36:13,235 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:13,235 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 16:36:13,235 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:13,235 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 16:36:13,235 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 5 +2024-05-14 16:36:13,235 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:13,235 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 16:36:13,235 DEBUG SenderThread:125128 [sender.py:send():378] send: summary +2024-05-14 16:36:13,236 INFO SenderThread:125128 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:36:13,236 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:13,236 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 16:36:13,236 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 6 +2024-05-14 16:36:13,236 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:13,236 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 16:36:13,236 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:13,236 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 16:36:13,238 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:36:13,305 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 7 +2024-05-14 16:36:13,306 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:13,306 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 16:36:13,306 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:13,306 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 16:36:14,103 INFO Thread-12 :125128 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/output.log +2024-05-14 16:36:14,103 INFO Thread-12 :125128 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/config.yaml +2024-05-14 16:36:14,103 INFO Thread-12 :125128 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/wandb-summary.json +2024-05-14 16:36:14,230 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:36:15,606 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 8 +2024-05-14 16:36:15,606 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:36:15,606 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:15,606 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 16:36:15,607 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:15,607 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 16:36:15,607 INFO SenderThread:125128 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 16:36:15,607 INFO SenderThread:125128 [job_builder.py:_get_source_type():576] no source found +2024-05-14 16:36:15,607 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 9 +2024-05-14 16:36:15,607 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:15,607 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 16:36:15,607 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:15,607 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 16:36:15,607 INFO SenderThread:125128 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 16:36:16,105 INFO SenderThread:125128 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/output.log +2024-05-14 16:36:16,105 INFO SenderThread:125128 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files +2024-05-14 16:36:16,105 INFO SenderThread:125128 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/config.yaml config.yaml +2024-05-14 16:36:16,105 INFO SenderThread:125128 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/requirements.txt requirements.txt +2024-05-14 16:36:16,106 INFO SenderThread:125128 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/wandb-metadata.json wandb-metadata.json +2024-05-14 16:36:16,106 INFO SenderThread:125128 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/output.log output.log +2024-05-14 16:36:16,107 INFO SenderThread:125128 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/wandb-summary.json wandb-summary.json +2024-05-14 16:36:16,107 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 10 +2024-05-14 16:36:16,108 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:16,108 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 16:36:16,110 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:16,110 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 16:36:16,110 INFO SenderThread:125128 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:36:16,230 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:36:16,230 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:36:16,331 INFO wandb-upload_0:125128 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/config.yaml +2024-05-14 16:36:16,503 INFO wandb-upload_1:125128 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/requirements.txt +2024-05-14 16:36:16,575 INFO wandb-upload_3:125128 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/wandb-summary.json +2024-05-14 16:36:16,606 INFO wandb-upload_2:125128 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/files/output.log +2024-05-14 16:36:16,807 INFO Thread-11 (_thread_body):125128 [sender.py:transition_state():613] send defer: 11 +2024-05-14 16:36:16,808 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:16,808 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 16:36:16,808 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:16,808 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 16:36:16,808 INFO SenderThread:125128 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 16:36:16,808 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 12 +2024-05-14 16:36:16,808 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:16,808 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 16:36:16,809 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:16,809 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 16:36:16,809 INFO SenderThread:125128 [file_stream.py:finish():601] file stream finish called +2024-05-14 16:36:17,017 INFO SenderThread:125128 [file_stream.py:finish():605] file stream finish is done +2024-05-14 16:36:17,017 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 13 +2024-05-14 16:36:17,017 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:17,017 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 16:36:17,018 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:17,018 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 16:36:17,018 INFO SenderThread:125128 [sender.py:transition_state():613] send defer: 14 +2024-05-14 16:36:17,018 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:36:17,018 DEBUG SenderThread:125128 [sender.py:send():378] send: final +2024-05-14 16:36:17,018 INFO HandlerThread:125128 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 16:36:17,018 DEBUG SenderThread:125128 [sender.py:send():378] send: footer +2024-05-14 16:36:17,018 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: defer +2024-05-14 16:36:17,018 INFO SenderThread:125128 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 16:36:17,019 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:36:17,019 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:36:17,019 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:36:17,019 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 16:36:17,019 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:36:17,020 DEBUG SenderThread:125128 [sender.py:send_request():405] send_request: server_info +2024-05-14 16:36:17,020 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 16:36:17,021 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 16:36:17,021 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 16:36:17,083 INFO MainThread:125128 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 16:36:17,084 INFO MainThread:125128 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 16:36:17,084 INFO MainThread:125128 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 16:36:17,084 DEBUG HandlerThread:125128 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 16:36:17,084 INFO HandlerThread:125128 [handler.py:finish():882] shutting down handler +2024-05-14 16:36:18,020 INFO WriterThread:125128 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/run-tnpwn56x.wandb +2024-05-14 16:36:18,083 INFO SenderThread:125128 [sender.py:finish():1545] shutting down sender +2024-05-14 16:36:18,083 INFO SenderThread:125128 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:36:18,083 INFO SenderThread:125128 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/logs/debug.log b/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..9b506ed8f541514cc19f5ea5b53cd45e4d701f7f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-14 16:36:06,853 INFO MainThread:123904 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_setup.py:_flush():76] Configure stats pid to 123904 +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_setup.py:_flush():76] Loading settings from /data/cronscript/lm-evaluation-harness/wandb/settings +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-14 16:36:06,854 WARNING MainThread:123904 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_init.py:_log_setup():520] Logging user logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/logs/debug.log +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_init.py:_log_setup():521] Logging internal logs to /data/cronscript/lm-evaluation-harness/wandb/run-20240514_163606-tnpwn56x/logs/debug-internal.log +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_init.py:init():560] calling init triggers +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_init.py:init():610] starting backend +2024-05-14 16:36:06,854 INFO MainThread:123904 [wandb_init.py:init():614] setting up manager +2024-05-14 16:36:06,855 INFO MainThread:123904 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-14 16:36:06,856 INFO MainThread:123904 [wandb_init.py:init():622] backend started and connected +2024-05-14 16:36:06,859 INFO MainThread:123904 [wandb_init.py:init():711] updated telemetry +2024-05-14 16:36:06,870 INFO MainThread:123904 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-14 16:36:07,106 INFO MainThread:123904 [wandb_run.py:_on_init():2396] communicating current version +2024-05-14 16:36:07,182 INFO MainThread:123904 [wandb_run.py:_on_init():2405] got version response +2024-05-14 16:36:07,183 INFO MainThread:123904 [wandb_init.py:init():795] starting run threads in backend +2024-05-14 16:36:07,389 INFO MainThread:123904 [wandb_run.py:_console_start():2374] atexit reg +2024-05-14 16:36:07,389 INFO MainThread:123904 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-14 16:36:07,389 INFO MainThread:123904 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-14 16:36:07,389 INFO MainThread:123904 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-14 16:36:07,394 INFO MainThread:123904 [wandb_init.py:init():838] run started, returning control to user process +2024-05-14 16:36:18,085 WARNING MsgRouterThr:123904 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/config.yaml b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..46cf6f00b3bd7716effe94e46d60c979920d0178 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.40.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1715705121 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.40.2 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..c38c906b3bb22119fb3818925f707320d2bc50e1 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log @@ -0,0 +1,33 @@ + +2024-05-14:16:45:22,510 INFO [__main__.py:251] Verbosity set to INFO +2024-05-14:16:45:27,174 INFO [__main__.py:335] Selected Tasks: ['indiccopa-hi'] +2024-05-14:16:45:27,176 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-14:16:45:27,177 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/data/cronscript/ckpts//hf_ckpt//global_step120'} +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/data/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/data/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/data/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/data/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/data/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 928, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 631, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 686, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 369, in cached_file + raise EnvironmentError( +OSError: /data/cronscript/ckpts//hf_ckpt//global_step120 does not appear to have a file named config.json. Checkout 'https://huggingface.co//data/cronscript/ckpts//hf_ckpt//global_step120/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..d63edd781bd5bdbb7f67523ac1ba9f0f1ed392dc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/requirements.txt @@ -0,0 +1,163 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.3 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.2 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.4 +aiosignal==1.3.1 +antlr4-python3-runtime==4.9.3 +anyio==4.3.0 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +distro==1.9.0 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.0 +expecttest==0.2.1 +filelock==3.13.4 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.62.1 +h11==0.14.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +httpcore==1.0.5 +httpx==0.27.0 +huggingface-hub==0.23.0 +identify==2.5.35 +idna==3.7 +importlib_resources==6.4.0 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.3.0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +omegaconf==2.3.0 +openai==1.29.0 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.10.0 +perfetto==0.7.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.0 +pluggy==1.4.0 +portalocker==2.8.2 +pre-commit==3.3.3 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.0.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycountry==23.12.11 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.1.1 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.2 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==1.5.0 +safetensors==0.4.3 +scikit-learn==1.4.2 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.1.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sniffio==1.3.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.3.2 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.2 +transformers==4.40.2 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.25.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..cc5af3c3c11d5da784fe65d2d3d88c350a105c88 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-metadata.json @@ -0,0 +1,810 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-14T16:45:22.370386", + "startedAt": "2024-05-14T16:45:21.927406", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120", + "--tasks", + "indiccopa-hi", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/data/cronscript/lm-evaluation-harness", + "host": "vizzhy-150-3", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 76, + "cpu_count_logical": 152, + "cpu_freq": { + "current": 3382.256710526316, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3287.34, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3287.353, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3286.819, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3291.515, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3279.752, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3273.163, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3293.407, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3299.998, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3291.544, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3290.282, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3285.038, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3291.469, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 866.4415092468262, + "used": 863.4300994873047 + } + }, + "memory": { + "total": 1007.5000267028809 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..74ae82ca002f26112178f0cd636ac5b92bf8e035 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 6}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..171d579e60954d5cd7b74ccc81285561e4c809f8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/logs/debug-internal.log @@ -0,0 +1,182 @@ +2024-05-14 16:45:21,939 INFO StreamThr :137244 [internal.py:wandb_internal():85] W&B internal server running at pid: 137244, started at: 2024-05-14 16:45:21.938990 +2024-05-14 16:45:21,942 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: status +2024-05-14 16:45:21,943 INFO WriterThread:137244 [datastore.py:open_for_write():87] open: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/run-grcnstg5.wandb +2024-05-14 16:45:21,944 DEBUG SenderThread:137244 [sender.py:send():378] send: header +2024-05-14 16:45:21,952 DEBUG SenderThread:137244 [sender.py:send():378] send: run +2024-05-14 16:45:22,182 INFO SenderThread:137244 [dir_watcher.py:__init__():211] watching files in: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files +2024-05-14 16:45:22,183 INFO SenderThread:137244 [sender.py:_start_run_threads():1123] run started: grcnstg5 with start time 1715705121.938422 +2024-05-14 16:45:22,190 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: check_version +2024-05-14 16:45:22,190 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: check_version +2024-05-14 16:45:22,279 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: run_start +2024-05-14 16:45:22,281 DEBUG HandlerThread:137244 [system_info.py:__init__():26] System info init +2024-05-14 16:45:22,281 DEBUG HandlerThread:137244 [system_info.py:__init__():41] System info init done +2024-05-14 16:45:22,281 INFO HandlerThread:137244 [system_monitor.py:start():194] Starting system monitor +2024-05-14 16:45:22,281 INFO SystemMonitor:137244 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-14 16:45:22,281 INFO HandlerThread:137244 [system_monitor.py:probe():214] Collecting system info +2024-05-14 16:45:22,281 INFO SystemMonitor:137244 [interfaces.py:start():188] Started cpu monitoring +2024-05-14 16:45:22,281 INFO SystemMonitor:137244 [interfaces.py:start():188] Started disk monitoring +2024-05-14 16:45:22,282 INFO SystemMonitor:137244 [interfaces.py:start():188] Started memory monitoring +2024-05-14 16:45:22,282 INFO SystemMonitor:137244 [interfaces.py:start():188] Started network monitoring +2024-05-14 16:45:22,370 DEBUG HandlerThread:137244 [system_info.py:probe():150] Probing system +2024-05-14 16:45:22,386 DEBUG HandlerThread:137244 [system_info.py:_probe_git():135] Probing git +2024-05-14 16:45:22,406 ERROR HandlerThread:137244 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/data/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /data/cronscript/lm-evaluation-harness' +2024-05-14 16:45:22,406 DEBUG HandlerThread:137244 [system_info.py:_probe_git():143] Probing git done +2024-05-14 16:45:22,406 DEBUG HandlerThread:137244 [system_info.py:probe():198] Probing system done +2024-05-14 16:45:22,406 DEBUG HandlerThread:137244 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-14T16:45:22.370386', 'startedAt': '2024-05-14T16:45:21.927406', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/data/cronscript/ckpts//hf_ckpt//global_step120', '--tasks', 'indiccopa-hi', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/data/cronscript/lm-evaluation-harness', 'host': 'vizzhy-150-3', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 76, 'cpu_count_logical': 152, 'cpu_freq': {'current': 3382.256710526316, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3287.34, 'min': 800.0, 'max': 3400.0}, {'current': 3287.353, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3286.819, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3291.515, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3279.752, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3273.163, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3293.407, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3299.998, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3291.544, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3290.282, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3285.038, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3291.469, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 866.4415092468262, 'used': 863.4300994873047}}, 'memory': {'total': 1007.5000267028809}} +2024-05-14 16:45:22,406 INFO HandlerThread:137244 [system_monitor.py:probe():224] Finished collecting system info +2024-05-14 16:45:22,406 INFO HandlerThread:137244 [system_monitor.py:probe():227] Publishing system info +2024-05-14 16:45:22,408 INFO HandlerThread:137244 [system_monitor.py:probe():229] Finished publishing system info +2024-05-14 16:45:22,411 DEBUG SenderThread:137244 [sender.py:send():378] send: files +2024-05-14 16:45:22,412 INFO SenderThread:137244 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-14 16:45:22,507 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: python_packages +2024-05-14 16:45:22,507 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: stop_status +2024-05-14 16:45:22,507 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: python_packages +2024-05-14 16:45:22,508 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: stop_status +2024-05-14 16:45:22,722 DEBUG SenderThread:137244 [sender.py:send():378] send: telemetry +2024-05-14 16:45:22,963 INFO wandb-upload_0:137244 [upload_job.py:push():130] Uploaded file /tmp/tmp94kc9y4ewandb/0qfkf8bx-wandb-metadata.json +2024-05-14 16:45:23,185 INFO Thread-12 :137244 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log +2024-05-14 16:45:23,185 INFO Thread-12 :137244 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/requirements.txt +2024-05-14 16:45:23,185 INFO Thread-12 :137244 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-metadata.json +2024-05-14 16:45:25,184 INFO Thread-12 :137244 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log +2024-05-14 16:45:27,176 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:45:28,344 DEBUG SenderThread:137244 [sender.py:send():378] send: exit +2024-05-14 16:45:28,344 INFO SenderThread:137244 [sender.py:send_exit():585] handling exit code: 1 +2024-05-14 16:45:28,344 INFO SenderThread:137244 [sender.py:send_exit():587] handling runtime: 6 +2024-05-14 16:45:28,345 INFO SenderThread:137244 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:45:28,345 INFO SenderThread:137244 [sender.py:send_exit():593] send defer +2024-05-14 16:45:28,346 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,346 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-14 16:45:28,346 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,346 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-14 16:45:28,346 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 1 +2024-05-14 16:45:28,346 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,346 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-14 16:45:28,346 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,346 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-14 16:45:28,346 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 2 +2024-05-14 16:45:28,346 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,346 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-14 16:45:28,346 INFO HandlerThread:137244 [system_monitor.py:finish():203] Stopping system monitor +2024-05-14 16:45:28,347 INFO HandlerThread:137244 [interfaces.py:finish():200] Joined cpu monitor +2024-05-14 16:45:28,347 DEBUG SystemMonitor:137244 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-14 16:45:28,347 INFO HandlerThread:137244 [interfaces.py:finish():200] Joined disk monitor +2024-05-14 16:45:28,347 DEBUG SystemMonitor:137244 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-14 16:45:28,347 INFO HandlerThread:137244 [interfaces.py:finish():200] Joined memory monitor +2024-05-14 16:45:28,347 DEBUG SystemMonitor:137244 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-14 16:45:28,347 INFO HandlerThread:137244 [interfaces.py:finish():200] Joined network monitor +2024-05-14 16:45:28,349 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,349 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-14 16:45:28,349 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 3 +2024-05-14 16:45:28,349 DEBUG SenderThread:137244 [sender.py:send():378] send: stats +2024-05-14 16:45:28,350 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,350 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-14 16:45:28,350 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,350 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-14 16:45:28,350 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 4 +2024-05-14 16:45:28,350 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,350 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-14 16:45:28,350 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,350 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-14 16:45:28,350 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 5 +2024-05-14 16:45:28,350 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,350 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-14 16:45:28,350 DEBUG SenderThread:137244 [sender.py:send():378] send: summary +2024-05-14 16:45:28,351 INFO SenderThread:137244 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-14 16:45:28,351 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,351 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-14 16:45:28,351 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 6 +2024-05-14 16:45:28,351 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,351 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-14 16:45:28,351 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,351 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-14 16:45:28,354 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: status_report +2024-05-14 16:45:28,426 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 7 +2024-05-14 16:45:28,426 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:28,426 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-14 16:45:28,427 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:28,427 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-14 16:45:29,187 INFO Thread-12 :137244 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/config.yaml +2024-05-14 16:45:29,187 INFO Thread-12 :137244 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log +2024-05-14 16:45:29,187 INFO Thread-12 :137244 [dir_watcher.py:_on_file_created():271] file/dir created: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-summary.json +2024-05-14 16:45:29,344 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:30,741 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 8 +2024-05-14 16:45:30,742 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:30,742 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:30,742 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-14 16:45:30,742 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:30,742 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-14 16:45:30,742 INFO SenderThread:137244 [job_builder.py:build():432] Attempting to build job artifact +2024-05-14 16:45:30,743 INFO SenderThread:137244 [job_builder.py:_get_source_type():576] no source found +2024-05-14 16:45:30,743 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 9 +2024-05-14 16:45:30,743 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:30,743 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-14 16:45:30,743 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:30,743 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-14 16:45:30,743 INFO SenderThread:137244 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-14 16:45:31,188 INFO SenderThread:137244 [dir_watcher.py:_on_file_modified():288] file/dir modified: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log +2024-05-14 16:45:31,188 INFO SenderThread:137244 [dir_watcher.py:finish():388] scan: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files +2024-05-14 16:45:31,188 INFO SenderThread:137244 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-summary.json wandb-summary.json +2024-05-14 16:45:31,188 INFO SenderThread:137244 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/config.yaml config.yaml +2024-05-14 16:45:31,189 INFO SenderThread:137244 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/requirements.txt requirements.txt +2024-05-14 16:45:31,189 INFO SenderThread:137244 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-metadata.json wandb-metadata.json +2024-05-14 16:45:31,189 INFO SenderThread:137244 [dir_watcher.py:finish():402] scan save: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log output.log +2024-05-14 16:45:31,189 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 10 +2024-05-14 16:45:31,191 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:31,192 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-14 16:45:31,194 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:31,194 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-14 16:45:31,194 INFO SenderThread:137244 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:45:31,345 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:31,345 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:31,433 INFO wandb-upload_1:137244 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/config.yaml +2024-05-14 16:45:31,602 INFO wandb-upload_0:137244 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/wandb-summary.json +2024-05-14 16:45:31,690 INFO wandb-upload_2:137244 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/requirements.txt +2024-05-14 16:45:31,704 INFO wandb-upload_3:137244 [upload_job.py:push():130] Uploaded file /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/files/output.log +2024-05-14 16:45:31,904 INFO Thread-11 (_thread_body):137244 [sender.py:transition_state():613] send defer: 11 +2024-05-14 16:45:31,904 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:31,904 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-14 16:45:31,905 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:31,905 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-14 16:45:31,905 INFO SenderThread:137244 [file_pusher.py:join():175] waiting for file pusher +2024-05-14 16:45:31,905 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 12 +2024-05-14 16:45:31,905 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:31,905 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-14 16:45:31,906 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:31,906 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-14 16:45:31,906 INFO SenderThread:137244 [file_stream.py:finish():601] file stream finish called +2024-05-14 16:45:32,127 INFO SenderThread:137244 [file_stream.py:finish():605] file stream finish is done +2024-05-14 16:45:32,127 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 13 +2024-05-14 16:45:32,127 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:32,127 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-14 16:45:32,127 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:32,127 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-14 16:45:32,127 INFO SenderThread:137244 [sender.py:transition_state():613] send defer: 14 +2024-05-14 16:45:32,127 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: defer +2024-05-14 16:45:32,127 INFO HandlerThread:137244 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-14 16:45:32,128 DEBUG SenderThread:137244 [sender.py:send():378] send: final +2024-05-14 16:45:32,128 DEBUG SenderThread:137244 [sender.py:send():378] send: footer +2024-05-14 16:45:32,128 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: defer +2024-05-14 16:45:32,128 INFO SenderThread:137244 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-14 16:45:32,128 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:32,128 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:32,129 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-14 16:45:32,129 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: server_info +2024-05-14 16:45:32,129 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: poll_exit +2024-05-14 16:45:32,129 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: get_summary +2024-05-14 16:45:32,129 DEBUG SenderThread:137244 [sender.py:send_request():405] send_request: server_info +2024-05-14 16:45:32,129 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-14 16:45:32,131 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-14 16:45:32,185 INFO MainThread:137244 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-14 16:45:32,185 INFO MainThread:137244 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-14 16:45:32,185 INFO MainThread:137244 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-14 16:45:32,185 DEBUG HandlerThread:137244 [handler.py:handle_request():158] handle_request: shutdown +2024-05-14 16:45:32,185 INFO HandlerThread:137244 [handler.py:finish():882] shutting down handler +2024-05-14 16:45:33,129 INFO WriterThread:137244 [datastore.py:close():296] close: /data/cronscript/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/run-grcnstg5.wandb +2024-05-14 16:45:33,185 INFO SenderThread:137244 [sender.py:finish():1545] shutting down sender +2024-05-14 16:45:33,185 INFO SenderThread:137244 [file_pusher.py:finish():169] shutting down file pusher +2024-05-14 16:45:33,185 INFO SenderThread:137244 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/run-grcnstg5.wandb b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/run-grcnstg5.wandb new file mode 100644 index 0000000000000000000000000000000000000000..81d13b17170efadfa78d04129bab0cb10669a3a3 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240514_164521-grcnstg5/run-grcnstg5.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..3d0b0ed1ccaec95dce0d431bdcbdd77d4ae737b5 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716396259 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..7791ceeab98ecda863b24f1d009175c4c80aba2b --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log @@ -0,0 +1,34 @@ + +2024-05-22:16:44:20,427 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:16:44:29,056 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:16:44:29,057 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:16:44:29,058 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100'} +2024-05-22:16:44:31,369 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..b3a9bb3f4b55f75bea905a25c70c8b917003d66c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T16:44:20.221778", + "startedAt": "2024-05-22T16:44:19.714725", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2337.5381500000003, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3218.818, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.012, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64118194580078 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..531a6481b5be5a00f5604a79cb8b6619a8887b9a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/logs/debug-internal.log @@ -0,0 +1,182 @@ +2024-05-22 16:44:19,735 INFO StreamThr :800 [internal.py:wandb_internal():85] W&B internal server running at pid: 800, started at: 2024-05-22 16:44:19.733213 +2024-05-22 16:44:19,739 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: status +2024-05-22 16:44:19,741 INFO WriterThread:800 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/run-a6yh7r45.wandb +2024-05-22 16:44:19,743 DEBUG SenderThread:800 [sender.py:send():378] send: header +2024-05-22 16:44:19,745 DEBUG SenderThread:800 [sender.py:send():378] send: run +2024-05-22 16:44:19,996 INFO SenderThread:800 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files +2024-05-22 16:44:19,996 INFO SenderThread:800 [sender.py:_start_run_threads():1123] run started: a6yh7r45 with start time 1716396259.733077 +2024-05-22 16:44:20,005 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 16:44:20,005 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: check_version +2024-05-22 16:44:20,123 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 16:44:20,126 DEBUG HandlerThread:800 [system_info.py:__init__():26] System info init +2024-05-22 16:44:20,126 DEBUG HandlerThread:800 [system_info.py:__init__():41] System info init done +2024-05-22 16:44:20,126 INFO HandlerThread:800 [system_monitor.py:start():194] Starting system monitor +2024-05-22 16:44:20,126 INFO SystemMonitor:800 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 16:44:20,126 INFO HandlerThread:800 [system_monitor.py:probe():214] Collecting system info +2024-05-22 16:44:20,134 INFO SystemMonitor:800 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 16:44:20,134 INFO SystemMonitor:800 [interfaces.py:start():188] Started disk monitoring +2024-05-22 16:44:20,140 INFO SystemMonitor:800 [interfaces.py:start():188] Started memory monitoring +2024-05-22 16:44:20,141 INFO SystemMonitor:800 [interfaces.py:start():188] Started network monitoring +2024-05-22 16:44:20,221 DEBUG HandlerThread:800 [system_info.py:probe():150] Probing system +2024-05-22 16:44:20,225 DEBUG HandlerThread:800 [system_info.py:_probe_git():135] Probing git +2024-05-22 16:44:20,235 ERROR HandlerThread:800 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 16:44:20,235 DEBUG HandlerThread:800 [system_info.py:_probe_git():143] Probing git done +2024-05-22 16:44:20,235 DEBUG HandlerThread:800 [system_info.py:probe():198] Probing system done +2024-05-22 16:44:20,235 DEBUG HandlerThread:800 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T16:44:20.221778', 'startedAt': '2024-05-22T16:44:19.714725', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step100', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2337.5381500000003, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3218.818, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.012, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64118194580078}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 16:44:20,236 INFO HandlerThread:800 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 16:44:20,236 INFO HandlerThread:800 [system_monitor.py:probe():227] Publishing system info +2024-05-22 16:44:20,239 INFO HandlerThread:800 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 16:44:20,244 DEBUG SenderThread:800 [sender.py:send():378] send: files +2024-05-22 16:44:20,244 INFO SenderThread:800 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 16:44:20,419 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 16:44:20,419 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: python_packages +2024-05-22 16:44:20,420 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 16:44:20,422 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: stop_status +2024-05-22 16:44:20,572 DEBUG SenderThread:800 [sender.py:send():378] send: telemetry +2024-05-22 16:44:20,813 INFO wandb-upload_0:800 [upload_job.py:push():130] Uploaded file /tmp/tmp__q6ioutwandb/jca97au8-wandb-metadata.json +2024-05-22 16:44:20,998 INFO Thread-12 :800 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/requirements.txt +2024-05-22 16:44:20,998 INFO Thread-12 :800 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log +2024-05-22 16:44:20,998 INFO Thread-12 :800 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-metadata.json +2024-05-22 16:44:22,998 INFO Thread-12 :800 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log +2024-05-22 16:44:25,574 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 16:44:31,005 INFO Thread-12 :800 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log +2024-05-22 16:44:31,058 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 16:44:31,381 DEBUG SenderThread:800 [sender.py:send():378] send: exit +2024-05-22 16:44:31,382 INFO SenderThread:800 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 16:44:31,382 INFO SenderThread:800 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 16:44:31,383 INFO SenderThread:800 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 16:44:31,383 INFO SenderThread:800 [sender.py:send_exit():593] send defer +2024-05-22 16:44:31,384 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:31,384 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 16:44:31,384 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:31,384 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 16:44:31,384 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 1 +2024-05-22 16:44:31,384 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:31,384 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 16:44:31,384 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:31,384 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 16:44:31,384 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 2 +2024-05-22 16:44:31,384 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:31,384 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 16:44:31,384 INFO HandlerThread:800 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 16:44:31,384 DEBUG SystemMonitor:800 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 16:44:31,384 DEBUG SystemMonitor:800 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 16:44:31,384 DEBUG SystemMonitor:800 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 16:44:31,385 INFO HandlerThread:800 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 16:44:31,385 INFO HandlerThread:800 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 16:44:31,385 INFO HandlerThread:800 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 16:44:31,386 INFO HandlerThread:800 [interfaces.py:finish():200] Joined network monitor +2024-05-22 16:44:31,386 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:31,386 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 16:44:31,386 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 3 +2024-05-22 16:44:31,386 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:31,386 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 16:44:31,386 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:31,386 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 16:44:31,386 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 4 +2024-05-22 16:44:31,386 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:31,386 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 16:44:31,386 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:31,386 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 16:44:31,386 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 5 +2024-05-22 16:44:31,386 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:31,387 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 16:44:31,387 DEBUG SenderThread:800 [sender.py:send():378] send: summary +2024-05-22 16:44:31,387 INFO SenderThread:800 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 16:44:31,388 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:31,388 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 16:44:31,388 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 6 +2024-05-22 16:44:31,388 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:31,388 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 16:44:31,388 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:31,388 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 16:44:31,393 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 16:44:31,485 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 7 +2024-05-22 16:44:31,485 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:31,485 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 16:44:31,485 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:31,485 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 16:44:32,007 INFO Thread-12 :800 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/config.yaml +2024-05-22 16:44:32,007 INFO Thread-12 :800 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-summary.json +2024-05-22 16:44:32,382 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 16:44:32,593 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 8 +2024-05-22 16:44:32,593 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 16:44:32,594 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:32,594 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 16:44:32,594 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:32,594 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 16:44:32,594 INFO SenderThread:800 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 16:44:32,594 INFO SenderThread:800 [job_builder.py:_get_source_type():576] no source found +2024-05-22 16:44:32,594 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 9 +2024-05-22 16:44:32,595 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:32,595 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 16:44:32,595 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:32,595 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 16:44:32,595 INFO SenderThread:800 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 16:44:33,008 INFO SenderThread:800 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log +2024-05-22 16:44:33,009 INFO SenderThread:800 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files +2024-05-22 16:44:33,009 INFO SenderThread:800 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log output.log +2024-05-22 16:44:33,009 INFO SenderThread:800 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-metadata.json wandb-metadata.json +2024-05-22 16:44:33,011 INFO SenderThread:800 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/config.yaml config.yaml +2024-05-22 16:44:33,011 INFO SenderThread:800 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/requirements.txt requirements.txt +2024-05-22 16:44:33,012 INFO SenderThread:800 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-summary.json wandb-summary.json +2024-05-22 16:44:33,012 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 10 +2024-05-22 16:44:33,012 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:33,012 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 16:44:33,012 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:33,012 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 16:44:33,012 INFO SenderThread:800 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 16:44:33,268 INFO wandb-upload_0:800 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/output.log +2024-05-22 16:44:33,382 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 16:44:33,382 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 16:44:33,596 INFO wandb-upload_2:800 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/requirements.txt +2024-05-22 16:44:33,652 INFO wandb-upload_1:800 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/config.yaml +2024-05-22 16:44:33,661 INFO wandb-upload_3:800 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/files/wandb-summary.json +2024-05-22 16:44:33,862 INFO Thread-11 (_thread_body):800 [sender.py:transition_state():613] send defer: 11 +2024-05-22 16:44:33,862 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:33,862 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 16:44:33,862 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:33,862 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 16:44:33,862 INFO SenderThread:800 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 16:44:33,862 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 12 +2024-05-22 16:44:33,862 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:33,862 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 16:44:33,863 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:33,863 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 16:44:33,863 INFO SenderThread:800 [file_stream.py:finish():601] file stream finish called +2024-05-22 16:44:33,939 INFO SenderThread:800 [file_stream.py:finish():605] file stream finish is done +2024-05-22 16:44:33,939 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 13 +2024-05-22 16:44:33,939 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:33,939 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 16:44:33,939 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:33,939 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 16:44:33,939 INFO SenderThread:800 [sender.py:transition_state():613] send defer: 14 +2024-05-22 16:44:33,940 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: defer +2024-05-22 16:44:33,940 INFO HandlerThread:800 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 16:44:33,940 DEBUG SenderThread:800 [sender.py:send():378] send: final +2024-05-22 16:44:33,940 DEBUG SenderThread:800 [sender.py:send():378] send: footer +2024-05-22 16:44:33,940 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: defer +2024-05-22 16:44:33,940 INFO SenderThread:800 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 16:44:33,941 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 16:44:33,941 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 16:44:33,941 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 16:44:33,941 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 16:44:33,941 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 16:44:33,941 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 16:44:33,941 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 16:44:33,941 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 16:44:33,941 DEBUG SenderThread:800 [sender.py:send_request():405] send_request: server_info +2024-05-22 16:44:33,994 INFO MainThread:800 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 16:44:33,994 INFO MainThread:800 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 16:44:33,994 INFO MainThread:800 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-22 16:44:33,995 DEBUG HandlerThread:800 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 16:44:33,995 INFO HandlerThread:800 [handler.py:finish():882] shutting down handler +2024-05-22 16:44:34,941 INFO WriterThread:800 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/run-a6yh7r45.wandb +2024-05-22 16:44:34,994 INFO SenderThread:800 [sender.py:finish():1545] shutting down sender +2024-05-22 16:44:34,994 INFO SenderThread:800 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 16:44:34,994 INFO SenderThread:800 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c055f9a0b8fe8c2a8a3f86149d16b96543d830af --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_setup.py:_flush():76] Configure stats pid to 644 +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 16:44:19,727 WARNING MainThread:644 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/logs/debug.log +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/logs/debug-internal.log +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_init.py:init():560] calling init triggers +2024-05-22 16:44:19,727 INFO MainThread:644 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 16:44:19,728 INFO MainThread:644 [wandb_init.py:init():610] starting backend +2024-05-22 16:44:19,728 INFO MainThread:644 [wandb_init.py:init():614] setting up manager +2024-05-22 16:44:19,731 INFO MainThread:644 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 16:44:19,732 INFO MainThread:644 [wandb_init.py:init():622] backend started and connected +2024-05-22 16:44:19,736 INFO MainThread:644 [wandb_init.py:init():711] updated telemetry +2024-05-22 16:44:19,744 INFO MainThread:644 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 16:44:20,005 INFO MainThread:644 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 16:44:20,117 INFO MainThread:644 [wandb_run.py:_on_init():2405] got version response +2024-05-22 16:44:20,117 INFO MainThread:644 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 16:44:20,420 INFO MainThread:644 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 16:44:20,420 INFO MainThread:644 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 16:44:20,421 INFO MainThread:644 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 16:44:20,421 INFO MainThread:644 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 16:44:20,424 INFO MainThread:644 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 16:44:34,996 WARNING MsgRouterThr:644 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/run-a6yh7r45.wandb b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/run-a6yh7r45.wandb new file mode 100644 index 0000000000000000000000000000000000000000..a2c5341cded9783e371da5c77ee2191ce1913f6a Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_164419-a6yh7r45/run-a6yh7r45.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c71c04afb9507c4aff36b572ce3df9dfd2478ac3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/config.yaml @@ -0,0 +1,32 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716399426 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/output.log b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..8b137891791fe96927ad78e64b0aad7bded08bdc --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/output.log @@ -0,0 +1 @@ + diff --git a/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..612b0796554200b1f3a93430106c389e70c7ea31 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T17:37:07.185554", + "startedAt": "2024-05-22T17:37:06.635155", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.49998125, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64346313476562 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d9018f15dee0f335bfb0a98e044ecf092caffb66 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/logs/debug-internal.log @@ -0,0 +1,46 @@ +2024-05-22 17:37:06,658 INFO StreamThr :1284 [internal.py:wandb_internal():85] W&B internal server running at pid: 1284, started at: 2024-05-22 17:37:06.654975 +2024-05-22 17:37:06,661 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: status +2024-05-22 17:37:06,662 INFO WriterThread:1284 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/run-ika7vp1p.wandb +2024-05-22 17:37:06,663 DEBUG SenderThread:1284 [sender.py:send():378] send: header +2024-05-22 17:37:06,667 DEBUG SenderThread:1284 [sender.py:send():378] send: run +2024-05-22 17:37:06,970 INFO SenderThread:1284 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files +2024-05-22 17:37:06,970 INFO SenderThread:1284 [sender.py:_start_run_threads():1123] run started: ika7vp1p with start time 1716399426.654831 +2024-05-22 17:37:06,975 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 17:37:06,975 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: check_version +2024-05-22 17:37:07,091 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 17:37:07,094 DEBUG HandlerThread:1284 [system_info.py:__init__():26] System info init +2024-05-22 17:37:07,094 DEBUG HandlerThread:1284 [system_info.py:__init__():41] System info init done +2024-05-22 17:37:07,094 INFO HandlerThread:1284 [system_monitor.py:start():194] Starting system monitor +2024-05-22 17:37:07,094 INFO SystemMonitor:1284 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 17:37:07,094 INFO HandlerThread:1284 [system_monitor.py:probe():214] Collecting system info +2024-05-22 17:37:07,101 INFO SystemMonitor:1284 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 17:37:07,101 INFO SystemMonitor:1284 [interfaces.py:start():188] Started disk monitoring +2024-05-22 17:37:07,101 INFO SystemMonitor:1284 [interfaces.py:start():188] Started memory monitoring +2024-05-22 17:37:07,102 INFO SystemMonitor:1284 [interfaces.py:start():188] Started network monitoring +2024-05-22 17:37:07,185 DEBUG HandlerThread:1284 [system_info.py:probe():150] Probing system +2024-05-22 17:37:07,188 DEBUG HandlerThread:1284 [system_info.py:_probe_git():135] Probing git +2024-05-22 17:37:07,198 ERROR HandlerThread:1284 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 17:37:07,198 DEBUG HandlerThread:1284 [system_info.py:_probe_git():143] Probing git done +2024-05-22 17:37:07,198 DEBUG HandlerThread:1284 [system_info.py:probe():198] Probing system done +2024-05-22 17:37:07,198 DEBUG HandlerThread:1284 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T17:37:07.185554', 'startedAt': '2024-05-22T17:37:06.635155', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step12000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.49998125, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64346313476562}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 17:37:07,198 INFO HandlerThread:1284 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 17:37:07,198 INFO HandlerThread:1284 [system_monitor.py:probe():227] Publishing system info +2024-05-22 17:37:07,202 INFO HandlerThread:1284 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 17:37:07,207 DEBUG SenderThread:1284 [sender.py:send():378] send: files +2024-05-22 17:37:07,207 INFO SenderThread:1284 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 17:37:07,416 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 17:37:07,417 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: python_packages +2024-05-22 17:37:07,418 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 17:37:07,419 DEBUG SenderThread:1284 [sender.py:send_request():405] send_request: stop_status +2024-05-22 17:37:07,552 DEBUG SenderThread:1284 [sender.py:send():378] send: telemetry +2024-05-22 17:37:07,779 INFO wandb-upload_0:1284 [upload_job.py:push():130] Uploaded file /tmp/tmp0xiwgsf9wandb/mgy2qut1-wandb-metadata.json +2024-05-22 17:37:07,972 INFO Thread-12 :1284 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/wandb-metadata.json +2024-05-22 17:37:07,973 INFO Thread-12 :1284 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/output.log +2024-05-22 17:37:07,973 INFO Thread-12 :1284 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/requirements.txt +2024-05-22 17:37:09,973 INFO Thread-12 :1284 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/files/output.log +2024-05-22 17:37:12,555 DEBUG HandlerThread:1284 [handler.py:handle_request():158] handle_request: status_report diff --git a/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..4bcdbb5669d41db9c2d94679a55dccda63bc5a5c --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/logs/debug.log @@ -0,0 +1,28 @@ +2024-05-22 17:37:06,648 INFO MainThread:1129 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_setup.py:_flush():76] Configure stats pid to 1129 +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 17:37:06,649 WARNING MainThread:1129 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/logs/debug.log +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/logs/debug-internal.log +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_init.py:init():560] calling init triggers +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_init.py:init():610] starting backend +2024-05-22 17:37:06,649 INFO MainThread:1129 [wandb_init.py:init():614] setting up manager +2024-05-22 17:37:06,653 INFO MainThread:1129 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 17:37:06,654 INFO MainThread:1129 [wandb_init.py:init():622] backend started and connected +2024-05-22 17:37:06,658 INFO MainThread:1129 [wandb_init.py:init():711] updated telemetry +2024-05-22 17:37:06,667 INFO MainThread:1129 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 17:37:06,974 INFO MainThread:1129 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 17:37:07,085 INFO MainThread:1129 [wandb_run.py:_on_init():2405] got version response +2024-05-22 17:37:07,085 INFO MainThread:1129 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 17:37:07,418 INFO MainThread:1129 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 17:37:07,420 INFO MainThread:1129 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 17:37:07,420 INFO MainThread:1129 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 17:37:07,421 INFO MainThread:1129 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 17:37:07,422 INFO MainThread:1129 [wandb_init.py:init():838] run started, returning control to user process diff --git a/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/run-ika7vp1p.wandb b/lm-evaluation-harness/wandb/run-20240522_173706-ika7vp1p/run-ika7vp1p.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/config.yaml b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..c87ecad6c35d9488a911ca33bc247ed37cda25ff --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.0 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716404207 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.0 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..574656bd831a3988eb3ac054aa7e3f8b4c1c0724 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log @@ -0,0 +1,34 @@ + +2024-05-22:18:56:48,598 INFO [__main__.py:251] Verbosity set to INFO +2024-05-22:18:56:57,036 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-22:18:56:57,037 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-22:18:56:57,037 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000'} +2024-05-22:18:56:59,369 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..a3573b3bbfd5b190fed4ccaed4ac2846002aec22 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.0 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.0 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..ef32b3488f77b86ebfb7ebfc037eadeb15226c0f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-22T18:56:48.397942", + "startedAt": "2024-05-22T18:56:47.860800", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2361.09520625, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3399.997, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.64026641845703 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..a097eeea77242eea0ef393ad3edfee34a2253d9a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-22 18:56:47,883 INFO StreamThr :2835 [internal.py:wandb_internal():85] W&B internal server running at pid: 2835, started at: 2024-05-22 18:56:47.881637 +2024-05-22 18:56:47,888 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: status +2024-05-22 18:56:47,888 INFO WriterThread:2835 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/run-tktaks6b.wandb +2024-05-22 18:56:47,890 DEBUG SenderThread:2835 [sender.py:send():378] send: header +2024-05-22 18:56:47,893 DEBUG SenderThread:2835 [sender.py:send():378] send: run +2024-05-22 18:56:48,189 INFO SenderThread:2835 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files +2024-05-22 18:56:48,189 INFO SenderThread:2835 [sender.py:_start_run_threads():1123] run started: tktaks6b with start time 1716404207.881495 +2024-05-22 18:56:48,198 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: check_version +2024-05-22 18:56:48,199 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: check_version +2024-05-22 18:56:48,320 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: run_start +2024-05-22 18:56:48,322 DEBUG HandlerThread:2835 [system_info.py:__init__():26] System info init +2024-05-22 18:56:48,322 DEBUG HandlerThread:2835 [system_info.py:__init__():41] System info init done +2024-05-22 18:56:48,322 INFO HandlerThread:2835 [system_monitor.py:start():194] Starting system monitor +2024-05-22 18:56:48,323 INFO SystemMonitor:2835 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-22 18:56:48,323 INFO HandlerThread:2835 [system_monitor.py:probe():214] Collecting system info +2024-05-22 18:56:48,329 INFO SystemMonitor:2835 [interfaces.py:start():188] Started cpu monitoring +2024-05-22 18:56:48,330 INFO SystemMonitor:2835 [interfaces.py:start():188] Started disk monitoring +2024-05-22 18:56:48,331 INFO SystemMonitor:2835 [interfaces.py:start():188] Started memory monitoring +2024-05-22 18:56:48,336 INFO SystemMonitor:2835 [interfaces.py:start():188] Started network monitoring +2024-05-22 18:56:48,397 DEBUG HandlerThread:2835 [system_info.py:probe():150] Probing system +2024-05-22 18:56:48,401 DEBUG HandlerThread:2835 [system_info.py:_probe_git():135] Probing git +2024-05-22 18:56:48,410 ERROR HandlerThread:2835 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-22 18:56:48,411 DEBUG HandlerThread:2835 [system_info.py:_probe_git():143] Probing git done +2024-05-22 18:56:48,411 DEBUG HandlerThread:2835 [system_info.py:probe():198] Probing system done +2024-05-22 18:56:48,411 DEBUG HandlerThread:2835 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-22T18:56:48.397942', 'startedAt': '2024-05-22T18:56:47.860800', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2361.09520625, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3399.997, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.64026641845703}}, 'memory': {'total': 1007.4379997253418}} +2024-05-22 18:56:48,411 INFO HandlerThread:2835 [system_monitor.py:probe():224] Finished collecting system info +2024-05-22 18:56:48,411 INFO HandlerThread:2835 [system_monitor.py:probe():227] Publishing system info +2024-05-22 18:56:48,414 INFO HandlerThread:2835 [system_monitor.py:probe():229] Finished publishing system info +2024-05-22 18:56:48,419 DEBUG SenderThread:2835 [sender.py:send():378] send: files +2024-05-22 18:56:48,419 INFO SenderThread:2835 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-22 18:56:48,592 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: python_packages +2024-05-22 18:56:48,592 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: python_packages +2024-05-22 18:56:48,594 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: stop_status +2024-05-22 18:56:48,595 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: stop_status +2024-05-22 18:56:48,718 DEBUG SenderThread:2835 [sender.py:send():378] send: telemetry +2024-05-22 18:56:48,969 INFO wandb-upload_0:2835 [upload_job.py:push():130] Uploaded file /tmp/tmp74i9ivz1wandb/kvhsaksq-wandb-metadata.json +2024-05-22 18:56:49,192 INFO Thread-12 :2835 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log +2024-05-22 18:56:49,192 INFO Thread-12 :2835 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/requirements.txt +2024-05-22 18:56:49,192 INFO Thread-12 :2835 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-metadata.json +2024-05-22 18:56:51,191 INFO Thread-12 :2835 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log +2024-05-22 18:56:53,723 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:56:59,038 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:56:59,198 INFO Thread-12 :2835 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log +2024-05-22 18:56:59,378 DEBUG SenderThread:2835 [sender.py:send():378] send: exit +2024-05-22 18:56:59,378 INFO SenderThread:2835 [sender.py:send_exit():585] handling exit code: 1 +2024-05-22 18:56:59,378 INFO SenderThread:2835 [sender.py:send_exit():587] handling runtime: 11 +2024-05-22 18:56:59,379 INFO SenderThread:2835 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:56:59,380 INFO SenderThread:2835 [sender.py:send_exit():593] send defer +2024-05-22 18:56:59,380 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:59,380 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-22 18:56:59,380 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:59,380 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-22 18:56:59,380 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 1 +2024-05-22 18:56:59,380 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:59,380 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-22 18:56:59,380 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:59,380 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-22 18:56:59,380 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 2 +2024-05-22 18:56:59,380 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:59,380 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-22 18:56:59,380 INFO HandlerThread:2835 [system_monitor.py:finish():203] Stopping system monitor +2024-05-22 18:56:59,381 DEBUG SystemMonitor:2835 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-22 18:56:59,381 DEBUG SystemMonitor:2835 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-22 18:56:59,381 DEBUG SystemMonitor:2835 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-22 18:56:59,383 INFO HandlerThread:2835 [interfaces.py:finish():200] Joined cpu monitor +2024-05-22 18:56:59,384 INFO HandlerThread:2835 [interfaces.py:finish():200] Joined disk monitor +2024-05-22 18:56:59,384 INFO HandlerThread:2835 [interfaces.py:finish():200] Joined memory monitor +2024-05-22 18:56:59,384 INFO HandlerThread:2835 [interfaces.py:finish():200] Joined network monitor +2024-05-22 18:56:59,384 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:59,384 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-22 18:56:59,384 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 3 +2024-05-22 18:56:59,384 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:59,384 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-22 18:56:59,385 DEBUG SenderThread:2835 [sender.py:send():378] send: stats +2024-05-22 18:56:59,386 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:59,386 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-22 18:56:59,386 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 4 +2024-05-22 18:56:59,386 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:59,386 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-22 18:56:59,386 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:59,386 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-22 18:56:59,386 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 5 +2024-05-22 18:56:59,386 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:59,386 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-22 18:56:59,386 DEBUG SenderThread:2835 [sender.py:send():378] send: summary +2024-05-22 18:56:59,387 INFO SenderThread:2835 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-22 18:56:59,387 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:59,387 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-22 18:56:59,387 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 6 +2024-05-22 18:56:59,387 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:59,387 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-22 18:56:59,387 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:59,388 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-22 18:56:59,392 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: status_report +2024-05-22 18:56:59,489 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 7 +2024-05-22 18:56:59,490 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:56:59,490 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-22 18:56:59,490 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:56:59,490 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-22 18:57:00,200 INFO Thread-12 :2835 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/config.yaml +2024-05-22 18:57:00,200 INFO Thread-12 :2835 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-summary.json +2024-05-22 18:57:00,378 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:57:00,744 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 8 +2024-05-22 18:57:00,744 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:57:00,744 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:57:00,744 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-22 18:57:00,744 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:57:00,744 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-22 18:57:00,744 INFO SenderThread:2835 [job_builder.py:build():432] Attempting to build job artifact +2024-05-22 18:57:00,745 INFO SenderThread:2835 [job_builder.py:_get_source_type():576] no source found +2024-05-22 18:57:00,745 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 9 +2024-05-22 18:57:00,745 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:57:00,745 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-22 18:57:00,745 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:57:00,745 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-22 18:57:00,745 INFO SenderThread:2835 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-22 18:57:01,201 INFO SenderThread:2835 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log +2024-05-22 18:57:01,201 INFO SenderThread:2835 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files +2024-05-22 18:57:01,202 INFO SenderThread:2835 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/requirements.txt requirements.txt +2024-05-22 18:57:01,202 INFO SenderThread:2835 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/config.yaml config.yaml +2024-05-22 18:57:01,204 INFO SenderThread:2835 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-metadata.json wandb-metadata.json +2024-05-22 18:57:01,204 INFO SenderThread:2835 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-summary.json wandb-summary.json +2024-05-22 18:57:01,205 INFO SenderThread:2835 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log output.log +2024-05-22 18:57:01,205 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 10 +2024-05-22 18:57:01,205 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:57:01,205 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-22 18:57:01,207 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:57:01,207 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-22 18:57:01,207 INFO SenderThread:2835 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:57:01,378 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:57:01,379 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:57:01,480 INFO wandb-upload_0:2835 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/requirements.txt +2024-05-22 18:57:01,774 INFO wandb-upload_1:2835 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/config.yaml +2024-05-22 18:57:01,789 INFO wandb-upload_2:2835 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/wandb-summary.json +2024-05-22 18:57:01,825 INFO wandb-upload_3:2835 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/files/output.log +2024-05-22 18:57:02,025 INFO Thread-11 (_thread_body):2835 [sender.py:transition_state():613] send defer: 11 +2024-05-22 18:57:02,025 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:57:02,025 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-22 18:57:02,026 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:57:02,026 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-22 18:57:02,026 INFO SenderThread:2835 [file_pusher.py:join():175] waiting for file pusher +2024-05-22 18:57:02,026 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 12 +2024-05-22 18:57:02,026 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:57:02,026 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-22 18:57:02,026 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:57:02,026 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-22 18:57:02,026 INFO SenderThread:2835 [file_stream.py:finish():601] file stream finish called +2024-05-22 18:57:02,108 INFO SenderThread:2835 [file_stream.py:finish():605] file stream finish is done +2024-05-22 18:57:02,108 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 13 +2024-05-22 18:57:02,108 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:57:02,109 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-22 18:57:02,109 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:57:02,109 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-22 18:57:02,109 INFO SenderThread:2835 [sender.py:transition_state():613] send defer: 14 +2024-05-22 18:57:02,109 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: defer +2024-05-22 18:57:02,109 INFO HandlerThread:2835 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-22 18:57:02,109 DEBUG SenderThread:2835 [sender.py:send():378] send: final +2024-05-22 18:57:02,109 DEBUG SenderThread:2835 [sender.py:send():378] send: footer +2024-05-22 18:57:02,109 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: defer +2024-05-22 18:57:02,109 INFO SenderThread:2835 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-22 18:57:02,110 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:57:02,110 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-22 18:57:02,110 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: server_info +2024-05-22 18:57:02,110 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: get_summary +2024-05-22 18:57:02,110 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-22 18:57:02,110 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-22 18:57:02,111 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:57:02,111 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: poll_exit +2024-05-22 18:57:02,111 DEBUG SenderThread:2835 [sender.py:send_request():405] send_request: server_info +2024-05-22 18:57:02,174 INFO MainThread:2835 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-22 18:57:02,174 INFO MainThread:2835 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-22 18:57:02,174 INFO MainThread:2835 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-22 18:57:02,174 DEBUG HandlerThread:2835 [handler.py:handle_request():158] handle_request: shutdown +2024-05-22 18:57:02,174 INFO HandlerThread:2835 [handler.py:finish():882] shutting down handler +2024-05-22 18:57:03,111 INFO WriterThread:2835 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/run-tktaks6b.wandb +2024-05-22 18:57:03,174 INFO SenderThread:2835 [sender.py:finish():1545] shutting down sender +2024-05-22 18:57:03,174 INFO SenderThread:2835 [file_pusher.py:finish():169] shutting down file pusher +2024-05-22 18:57:03,174 INFO SenderThread:2835 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/logs/debug.log b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..497e42b94e7b4d728b2d18ce7eb0ea4c09882b2d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-22 18:56:47,875 INFO MainThread:2680 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-22 18:56:47,875 INFO MainThread:2680 [wandb_setup.py:_flush():76] Configure stats pid to 2680 +2024-05-22 18:56:47,875 INFO MainThread:2680 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-22 18:56:47,875 INFO MainThread:2680 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-22 18:56:47,876 WARNING MainThread:2680 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/logs/debug.log +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/logs/debug-internal.log +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_init.py:init():560] calling init triggers +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_init.py:init():610] starting backend +2024-05-22 18:56:47,876 INFO MainThread:2680 [wandb_init.py:init():614] setting up manager +2024-05-22 18:56:47,880 INFO MainThread:2680 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-22 18:56:47,881 INFO MainThread:2680 [wandb_init.py:init():622] backend started and connected +2024-05-22 18:56:47,884 INFO MainThread:2680 [wandb_init.py:init():711] updated telemetry +2024-05-22 18:56:47,892 INFO MainThread:2680 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-22 18:56:48,198 INFO MainThread:2680 [wandb_run.py:_on_init():2396] communicating current version +2024-05-22 18:56:48,314 INFO MainThread:2680 [wandb_run.py:_on_init():2405] got version response +2024-05-22 18:56:48,314 INFO MainThread:2680 [wandb_init.py:init():795] starting run threads in backend +2024-05-22 18:56:48,593 INFO MainThread:2680 [wandb_run.py:_console_start():2374] atexit reg +2024-05-22 18:56:48,593 INFO MainThread:2680 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-22 18:56:48,593 INFO MainThread:2680 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-22 18:56:48,593 INFO MainThread:2680 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-22 18:56:48,596 INFO MainThread:2680 [wandb_init.py:init():838] run started, returning control to user process +2024-05-22 18:57:03,175 WARNING MsgRouterThr:2680 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/run-tktaks6b.wandb b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/run-tktaks6b.wandb new file mode 100644 index 0000000000000000000000000000000000000000..1cf7c16f720ff86273fe393425b4cd082e3ac9ab Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240522_185647-tktaks6b/run-tktaks6b.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..5b220684accfe87fb6c4747342140d7b751e1c73 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716441601 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/output.log b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..c385b292416935ed931c8a470f2707a516e0669a --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:05:20:02,067 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:05:20:10,573 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:05:20:10,574 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:05:20:10,574 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000'} +2024-05-23:05:20:12,904 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..6cacd5f3910710b7820532a2556f3a07820487dd --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T05:20:01.864522", + "startedAt": "2024-05-23T05:20:01.359120", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step4000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2322.112, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3274.578, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3359.564, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 212.18757247924805 + } + }, + "memory": { + "total": 1007.43798828125 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052001-u5aebb4f/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..32e2e9099c7792554c2589bda029fcad4ff832ec --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716441859 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..64f35d94f5af7b45cf9080fe8d15df8e8f101060 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:05:24:20,588 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:05:24:29,060 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:05:24:29,061 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:05:24:29,062 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000'} +2024-05-23:05:24:31,387 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..8150356038c46ec25f623f6e945d6dcb66a2e717 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.2.1 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..926992eb9984ac13182173bd9913b51dc429c372 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T05:24:20.383932", + "startedAt": "2024-05-23T05:24:19.903484", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2367.0835249999996, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 212.19213485717773 + } + }, + "memory": { + "total": 1007.43798828125 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..242fe2a753275f14c9c5b4b4f3d9fe60c0f9f901 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 05:24:19,925 INFO StreamThr :11379 [internal.py:wandb_internal():85] W&B internal server running at pid: 11379, started at: 2024-05-23 05:24:19.923389 +2024-05-23 05:24:19,930 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: status +2024-05-23 05:24:19,930 INFO WriterThread:11379 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/run-ibu2jsen.wandb +2024-05-23 05:24:19,932 DEBUG SenderThread:11379 [sender.py:send():378] send: header +2024-05-23 05:24:19,935 DEBUG SenderThread:11379 [sender.py:send():378] send: run +2024-05-23 05:24:20,186 INFO SenderThread:11379 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files +2024-05-23 05:24:20,186 INFO SenderThread:11379 [sender.py:_start_run_threads():1123] run started: ibu2jsen with start time 1716441859.923476 +2024-05-23 05:24:20,190 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 05:24:20,190 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: check_version +2024-05-23 05:24:20,310 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 05:24:20,312 DEBUG HandlerThread:11379 [system_info.py:__init__():26] System info init +2024-05-23 05:24:20,313 DEBUG HandlerThread:11379 [system_info.py:__init__():41] System info init done +2024-05-23 05:24:20,313 INFO HandlerThread:11379 [system_monitor.py:start():194] Starting system monitor +2024-05-23 05:24:20,313 INFO SystemMonitor:11379 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 05:24:20,313 INFO HandlerThread:11379 [system_monitor.py:probe():214] Collecting system info +2024-05-23 05:24:20,319 INFO SystemMonitor:11379 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 05:24:20,325 INFO SystemMonitor:11379 [interfaces.py:start():188] Started disk monitoring +2024-05-23 05:24:20,326 INFO SystemMonitor:11379 [interfaces.py:start():188] Started memory monitoring +2024-05-23 05:24:20,326 INFO SystemMonitor:11379 [interfaces.py:start():188] Started network monitoring +2024-05-23 05:24:20,383 DEBUG HandlerThread:11379 [system_info.py:probe():150] Probing system +2024-05-23 05:24:20,387 DEBUG HandlerThread:11379 [system_info.py:_probe_git():135] Probing git +2024-05-23 05:24:20,397 ERROR HandlerThread:11379 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 05:24:20,397 DEBUG HandlerThread:11379 [system_info.py:_probe_git():143] Probing git done +2024-05-23 05:24:20,397 DEBUG HandlerThread:11379 [system_info.py:probe():198] Probing system done +2024-05-23 05:24:20,397 DEBUG HandlerThread:11379 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T05:24:20.383932', 'startedAt': '2024-05-23T05:24:19.903484', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step8000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2367.0835249999996, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 212.19213485717773}}, 'memory': {'total': 1007.43798828125}} +2024-05-23 05:24:20,397 INFO HandlerThread:11379 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 05:24:20,397 INFO HandlerThread:11379 [system_monitor.py:probe():227] Publishing system info +2024-05-23 05:24:20,400 INFO HandlerThread:11379 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 05:24:20,405 DEBUG SenderThread:11379 [sender.py:send():378] send: files +2024-05-23 05:24:20,405 INFO SenderThread:11379 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 05:24:20,582 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 05:24:20,582 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: python_packages +2024-05-23 05:24:20,583 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 05:24:20,584 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: stop_status +2024-05-23 05:24:20,716 DEBUG SenderThread:11379 [sender.py:send():378] send: telemetry +2024-05-23 05:24:21,019 INFO wandb-upload_0:11379 [upload_job.py:push():130] Uploaded file /tmp/tmpqy9c_t51wandb/8n4wov62-wandb-metadata.json +2024-05-23 05:24:21,188 INFO Thread-12 :11379 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-metadata.json +2024-05-23 05:24:21,189 INFO Thread-12 :11379 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log +2024-05-23 05:24:21,189 INFO Thread-12 :11379 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/requirements.txt +2024-05-23 05:24:23,188 INFO Thread-12 :11379 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log +2024-05-23 05:24:25,718 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:24:31,062 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:24:31,196 INFO Thread-12 :11379 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log +2024-05-23 05:24:31,399 DEBUG SenderThread:11379 [sender.py:send():378] send: exit +2024-05-23 05:24:31,399 INFO SenderThread:11379 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 05:24:31,399 INFO SenderThread:11379 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 05:24:31,400 INFO SenderThread:11379 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:24:31,400 INFO SenderThread:11379 [sender.py:send_exit():593] send defer +2024-05-23 05:24:31,400 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:31,400 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 05:24:31,400 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:31,400 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 05:24:31,400 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 1 +2024-05-23 05:24:31,401 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:31,401 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 05:24:31,401 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:31,401 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 05:24:31,401 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 2 +2024-05-23 05:24:31,401 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:31,401 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 05:24:31,401 INFO HandlerThread:11379 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 05:24:31,401 DEBUG SystemMonitor:11379 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 05:24:31,402 DEBUG SystemMonitor:11379 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 05:24:31,402 INFO HandlerThread:11379 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 05:24:31,402 DEBUG SystemMonitor:11379 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 05:24:31,402 INFO HandlerThread:11379 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 05:24:31,403 INFO HandlerThread:11379 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 05:24:31,404 INFO HandlerThread:11379 [interfaces.py:finish():200] Joined network monitor +2024-05-23 05:24:31,404 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:31,404 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 05:24:31,404 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 3 +2024-05-23 05:24:31,404 DEBUG SenderThread:11379 [sender.py:send():378] send: stats +2024-05-23 05:24:31,405 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:31,405 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 05:24:31,406 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:31,406 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 05:24:31,406 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 4 +2024-05-23 05:24:31,406 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:31,406 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 05:24:31,406 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:31,406 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 05:24:31,406 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 5 +2024-05-23 05:24:31,406 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:31,406 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 05:24:31,406 DEBUG SenderThread:11379 [sender.py:send():378] send: summary +2024-05-23 05:24:31,407 INFO SenderThread:11379 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 05:24:31,407 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:31,407 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 05:24:31,407 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 6 +2024-05-23 05:24:31,407 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:31,407 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 05:24:31,407 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:31,407 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 05:24:31,412 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 05:24:31,477 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 7 +2024-05-23 05:24:31,477 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:31,477 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 05:24:31,477 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:31,477 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 05:24:32,198 INFO Thread-12 :11379 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/config.yaml +2024-05-23 05:24:32,198 INFO Thread-12 :11379 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-summary.json +2024-05-23 05:24:32,399 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:24:32,741 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 8 +2024-05-23 05:24:32,741 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:24:32,741 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:32,741 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 05:24:32,741 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:32,741 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 05:24:32,741 INFO SenderThread:11379 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 05:24:32,742 INFO SenderThread:11379 [job_builder.py:_get_source_type():576] no source found +2024-05-23 05:24:32,742 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 9 +2024-05-23 05:24:32,742 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:32,742 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 05:24:32,742 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:32,742 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 05:24:32,742 INFO SenderThread:11379 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 05:24:33,199 INFO SenderThread:11379 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log +2024-05-23 05:24:33,199 INFO SenderThread:11379 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files +2024-05-23 05:24:33,199 INFO SenderThread:11379 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/requirements.txt requirements.txt +2024-05-23 05:24:33,200 INFO SenderThread:11379 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-summary.json wandb-summary.json +2024-05-23 05:24:33,202 INFO SenderThread:11379 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log output.log +2024-05-23 05:24:33,202 INFO SenderThread:11379 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-metadata.json wandb-metadata.json +2024-05-23 05:24:33,202 INFO SenderThread:11379 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/config.yaml config.yaml +2024-05-23 05:24:33,202 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 10 +2024-05-23 05:24:33,202 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:33,203 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 05:24:33,203 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:33,203 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 05:24:33,203 INFO SenderThread:11379 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:24:33,399 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:24:33,399 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:24:33,444 INFO wandb-upload_0:11379 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/requirements.txt +2024-05-23 05:24:33,839 INFO wandb-upload_1:11379 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/wandb-summary.json +2024-05-23 05:24:33,915 INFO wandb-upload_3:11379 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/config.yaml +2024-05-23 05:24:34,070 INFO wandb-upload_2:11379 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/files/output.log +2024-05-23 05:24:34,270 INFO Thread-11 (_thread_body):11379 [sender.py:transition_state():613] send defer: 11 +2024-05-23 05:24:34,270 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:34,270 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 05:24:34,270 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:34,271 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 05:24:34,271 INFO SenderThread:11379 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 05:24:34,271 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 12 +2024-05-23 05:24:34,271 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:34,271 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 05:24:34,271 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:34,271 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 05:24:34,271 INFO SenderThread:11379 [file_stream.py:finish():601] file stream finish called +2024-05-23 05:24:34,340 INFO SenderThread:11379 [file_stream.py:finish():605] file stream finish is done +2024-05-23 05:24:34,340 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 13 +2024-05-23 05:24:34,340 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:34,340 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 05:24:34,340 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:34,340 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 05:24:34,341 INFO SenderThread:11379 [sender.py:transition_state():613] send defer: 14 +2024-05-23 05:24:34,341 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: defer +2024-05-23 05:24:34,341 INFO HandlerThread:11379 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 05:24:34,341 DEBUG SenderThread:11379 [sender.py:send():378] send: final +2024-05-23 05:24:34,341 DEBUG SenderThread:11379 [sender.py:send():378] send: footer +2024-05-23 05:24:34,341 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: defer +2024-05-23 05:24:34,341 INFO SenderThread:11379 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 05:24:34,342 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:24:34,342 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 05:24:34,342 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 05:24:34,342 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 05:24:34,342 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 05:24:34,342 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 05:24:34,342 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:24:34,342 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 05:24:34,343 DEBUG SenderThread:11379 [sender.py:send_request():405] send_request: server_info +2024-05-23 05:24:34,395 INFO MainThread:11379 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 05:24:34,395 INFO MainThread:11379 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 05:24:34,395 INFO MainThread:11379 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 05:24:34,395 DEBUG HandlerThread:11379 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 05:24:34,395 INFO HandlerThread:11379 [handler.py:finish():882] shutting down handler +2024-05-23 05:24:35,342 INFO WriterThread:11379 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/run-ibu2jsen.wandb +2024-05-23 05:24:35,395 INFO SenderThread:11379 [sender.py:finish():1545] shutting down sender +2024-05-23 05:24:35,395 INFO SenderThread:11379 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 05:24:35,395 INFO SenderThread:11379 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7250974072fa6c5eb13b7fe60956e6fccabd7100 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 05:24:19,917 INFO MainThread:11224 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 05:24:19,917 INFO MainThread:11224 [wandb_setup.py:_flush():76] Configure stats pid to 11224 +2024-05-23 05:24:19,917 INFO MainThread:11224 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 05:24:19,917 INFO MainThread:11224 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 05:24:19,917 INFO MainThread:11224 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 05:24:19,918 WARNING MainThread:11224 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/logs/debug.log +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/logs/debug-internal.log +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_init.py:init():560] calling init triggers +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_init.py:init():610] starting backend +2024-05-23 05:24:19,918 INFO MainThread:11224 [wandb_init.py:init():614] setting up manager +2024-05-23 05:24:19,922 INFO MainThread:11224 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 05:24:19,923 INFO MainThread:11224 [wandb_init.py:init():622] backend started and connected +2024-05-23 05:24:19,926 INFO MainThread:11224 [wandb_init.py:init():711] updated telemetry +2024-05-23 05:24:19,934 INFO MainThread:11224 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 05:24:20,190 INFO MainThread:11224 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 05:24:20,304 INFO MainThread:11224 [wandb_run.py:_on_init():2405] got version response +2024-05-23 05:24:20,304 INFO MainThread:11224 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 05:24:20,584 INFO MainThread:11224 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 05:24:20,584 INFO MainThread:11224 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 05:24:20,584 INFO MainThread:11224 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 05:24:20,584 INFO MainThread:11224 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 05:24:20,586 INFO MainThread:11224 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 05:24:35,396 WARNING MsgRouterThr:11224 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/run-ibu2jsen.wandb b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/run-ibu2jsen.wandb new file mode 100644 index 0000000000000000000000000000000000000000..42d1ab4fc62b149623260dd09aa5a132d37a53dc Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_052419-ibu2jsen/run-ibu2jsen.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/config.yaml b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..35f76cad639bceae9e576d46f7aba2132880157d --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/config.yaml @@ -0,0 +1,43 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.41.1 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1716467682 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 23 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.41.1 + 8: + - 5 + 13: linux-x86_64 diff --git a/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..2ea66be298b0fd575f4364019caf80c4f321c4b6 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log @@ -0,0 +1,34 @@ + +2024-05-23:12:34:43,459 INFO [__main__.py:251] Verbosity set to INFO +2024-05-23:12:34:51,934 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'hellaswag', 'mrpc', 'openbookqa', 'sst2', 'winogrande'] +2024-05-23:12:34:51,935 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-23:12:34:51,935 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000'} +2024-05-23:12:34:54,227 INFO [huggingface.py:164] Using device 'cuda' +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 190, in __init__ + self._get_config( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 471, in _get_config + self._config = transformers.AutoConfig.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/configuration_auto.py", line 934, in from_pretrained + config_dict, unused_kwargs = PretrainedConfig.get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 632, in get_config_dict + config_dict, kwargs = cls._get_config_dict(pretrained_model_name_or_path, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/configuration_utils.py", line 689, in _get_config_dict + resolved_config_file = cached_file( + File "/usr/local/lib/python3.10/dist-packages/transformers/utils/hub.py", line 370, in cached_file + raise EnvironmentError( +OSError: /mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000 does not appear to have a file named config.json. Checkout 'https://huggingface.co//mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000/tree/main' for available files. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/requirements.txt b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..f675c3016b5332c1acf28f436e0b60adeead9c12 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/requirements.txt @@ -0,0 +1,155 @@ +DataProperty==1.0.1 +GitPython==3.1.43 +Jinja2==3.1.4 +Markdown==3.6 +MarkupSafe==2.1.5 +Pillow-SIMD==7.0.0.post3 +PyYAML==6.0 +Werkzeug==3.0.3 +absl-py==2.1.0 +accelerate==0.30.1 +aiohttp==3.9.5 +aiosignal==1.3.1 +async-timeout==4.0.3 +attrs==23.2.0 +av==9.2.0 +cachetools==5.3.3 +certifi==2024.2.2 +cffi==1.15.1 +cfgv==3.4.0 +chardet==5.2.0 +charset-normalizer==3.3.2 +click==8.1.7 +cmake==3.29.2 +colorama==0.4.6 +datasets==2.19.1 +deepspeed==0.12.4+hpu.synapse.v1.15.1 +dill==0.3.8 +distlib==0.3.8 +docker-pycreds==0.4.0 +einops==0.8.0 +evaluate==0.4.2 +exceptiongroup==1.2.1 +expecttest==0.2.1 +filelock==3.14.0 +frozenlist==1.4.1 +fsspec==2024.3.1 +gitdb==4.0.11 +google-auth-oauthlib==0.4.6 +google-auth==2.29.0 +grpcio==1.63.0 +habana-media-loader==1.15.1.15 +habana-pyhlml==1.15.1.15 +habana-torch-dataloader==1.15.1.15 +habana-torch-plugin==1.15.1.15 +habana_gpu_migration==1.15.1.15 +habana_quantization_toolkit==1.15.1.15 +hjson==3.1.0 +huggingface-hub==0.23.1 +identify==2.5.36 +idna==3.7 +iniconfig==2.0.0 +joblib==1.4.2 +jsonlines==4.0.0 +lightning-habana==1.4.0 +lightning-utilities==0.11.2 +lightning==2.2.0.post0 +lm_eval==0.4.2 +lm_eval==0.4.2 +lm_eval==0.4.2 +lxml==5.2.2 +mbstrdecoder==1.1.3 +more-itertools==10.2.0 +mpi4py==3.1.4 +mpmath==1.3.0 +multidict==6.0.5 +multiprocess==0.70.16 +networkx==3.3 +ninja==1.11.1.1 +nltk==3.8.1 +nodeenv==1.8.0 +numexpr==2.10.0 +numpy==1.23.5 +oauthlib==3.2.2 +packaging==24.0 +pandas==2.0.1 +pathspec==0.12.1 +pathvalidate==3.2.0 +peft==0.11.1 +perfetto==0.7.0 +pillow==10.3.0 +pip==22.0.2 +pip==23.3.1 +platformdirs==4.2.1 +pluggy==1.5.0 +portalocker==2.8.2 +pre-commit==3.3.3 +pretty-errors==1.2.25 +protobuf==3.20.3 +psutil==5.9.8 +py-cpuinfo==9.0.0 +pyarrow-hotfix==0.6 +pyarrow==16.1.0 +pyasn1==0.6.0 +pyasn1_modules==0.4.0 +pybind11==2.10.4 +pycparser==2.22 +pydantic==1.10.13 +pynvml==8.0.4 +pytablewriter==1.2.0 +pytest==8.2.0 +python-dateutil==2.9.0.post0 +pytorch-lightning==2.2.4 +pytz==2024.1 +regex==2023.5.5 +requests-oauthlib==2.0.0 +requests==2.31.0 +rouge_score==0.1.2 +rsa==4.9 +sacrebleu==2.4.2 +safetensors==0.4.3 +scikit-learn==1.5.0 +scipy==1.13.1 +sentencepiece==0.2.0 +sentry-sdk==2.3.0 +setproctitle==1.3.3 +setuptools==59.6.0 +setuptools==69.5.1 +six==1.16.0 +smmap==5.0.1 +sqlitedict==2.1.0 +symengine==0.11.0 +sympy==1.12 +tabledata==1.3.3 +tabulate==0.9.0 +tcolorpy==0.1.6 +tdqm==0.0.1 +tensorboard-data-server==0.6.1 +tensorboard-plugin-wit==1.8.1 +tensorboard==2.11.2 +threadpoolctl==3.5.0 +tokenizers==0.19.1 +tomli==2.0.1 +torch==2.2.0a0+git8964477 +torch_tb_profiler==0.4.0 +torchaudio==2.2.0+08901ad +torchdata==0.7.1+5e6f7b7 +torchmetrics==1.4.0 +torchtext==0.17.0+400da5c +torchvision==0.17.0+b2383d4 +tqdm-multiprocess==0.0.11 +tqdm==4.66.4 +transformers==4.41.1 +typepy==1.3.2 +typing_extensions==4.11.0 +tzdata==2024.1 +urllib3==1.26.18 +virtualenv==20.26.1 +wandb==0.17.0 +wheel==0.37.1 +wheel==0.43.0 +word2number==1.1 +xxhash==3.4.1 +yamllint==1.35.1 +yarl==1.9.4 +zstandard==0.22.0 \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..dac6140ac2ce12394fa4a4d936f06cfcd5cb9e6f --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-23T12:34:43.249816", + "startedAt": "2024-05-23T12:34:42.496944", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000", + "--tasks", + "hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc", + "--batch_size", + "auto", + "--wandb_args", + "project=bharatgpt,group=trial_expt_2" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2334.2853812500002, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3399.998, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 209.58416366577148 + } + }, + "memory": { + "total": 1007.4379425048828 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-summary.json b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..8bf99d152ad35c3699ec8600ecb8b169d4e35875 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb": {"runtime": 11}} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..5bbec9b7b6cca80505f0449ac292728a6f7cf924 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/logs/debug-internal.log @@ -0,0 +1,183 @@ +2024-05-23 12:34:42,520 INFO StreamThr :2853 [internal.py:wandb_internal():85] W&B internal server running at pid: 2853, started at: 2024-05-23 12:34:42.517062 +2024-05-23 12:34:42,523 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: status +2024-05-23 12:34:42,524 INFO WriterThread:2853 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/run-rpj1bedw.wandb +2024-05-23 12:34:42,529 DEBUG SenderThread:2853 [sender.py:send():378] send: header +2024-05-23 12:34:42,529 DEBUG SenderThread:2853 [sender.py:send():378] send: run +2024-05-23 12:34:42,824 INFO SenderThread:2853 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files +2024-05-23 12:34:42,825 INFO SenderThread:2853 [sender.py:_start_run_threads():1123] run started: rpj1bedw with start time 1716467682.516922 +2024-05-23 12:34:42,829 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: check_version +2024-05-23 12:34:42,829 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: check_version +2024-05-23 12:34:43,154 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: run_start +2024-05-23 12:34:43,157 DEBUG HandlerThread:2853 [system_info.py:__init__():26] System info init +2024-05-23 12:34:43,157 DEBUG HandlerThread:2853 [system_info.py:__init__():41] System info init done +2024-05-23 12:34:43,157 INFO HandlerThread:2853 [system_monitor.py:start():194] Starting system monitor +2024-05-23 12:34:43,157 INFO SystemMonitor:2853 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-23 12:34:43,157 INFO HandlerThread:2853 [system_monitor.py:probe():214] Collecting system info +2024-05-23 12:34:43,164 INFO SystemMonitor:2853 [interfaces.py:start():188] Started cpu monitoring +2024-05-23 12:34:43,169 INFO SystemMonitor:2853 [interfaces.py:start():188] Started disk monitoring +2024-05-23 12:34:43,170 INFO SystemMonitor:2853 [interfaces.py:start():188] Started memory monitoring +2024-05-23 12:34:43,173 INFO SystemMonitor:2853 [interfaces.py:start():188] Started network monitoring +2024-05-23 12:34:43,249 DEBUG HandlerThread:2853 [system_info.py:probe():150] Probing system +2024-05-23 12:34:43,253 DEBUG HandlerThread:2853 [system_info.py:_probe_git():135] Probing git +2024-05-23 12:34:43,263 ERROR HandlerThread:2853 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-23 12:34:43,263 DEBUG HandlerThread:2853 [system_info.py:_probe_git():143] Probing git done +2024-05-23 12:34:43,263 DEBUG HandlerThread:2853 [system_info.py:probe():198] Probing system done +2024-05-23 12:34:43,263 DEBUG HandlerThread:2853 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-23T12:34:43.249816', 'startedAt': '2024-05-23T12:34:42.496944', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/checkpoint/llamav2-3b//hf_ckpt//global_step22000', '--tasks', 'hellaswag,arc_easy,openbookqa,winogrande,sst2,mrpc', '--batch_size', 'auto', '--wandb_args', 'project=bharatgpt,group=trial_expt_2'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2334.2853812500002, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3399.998, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 209.58416366577148}}, 'memory': {'total': 1007.4379425048828}} +2024-05-23 12:34:43,263 INFO HandlerThread:2853 [system_monitor.py:probe():224] Finished collecting system info +2024-05-23 12:34:43,263 INFO HandlerThread:2853 [system_monitor.py:probe():227] Publishing system info +2024-05-23 12:34:43,266 INFO HandlerThread:2853 [system_monitor.py:probe():229] Finished publishing system info +2024-05-23 12:34:43,271 DEBUG SenderThread:2853 [sender.py:send():378] send: files +2024-05-23 12:34:43,272 INFO SenderThread:2853 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-23 12:34:43,453 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: python_packages +2024-05-23 12:34:43,453 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: python_packages +2024-05-23 12:34:43,455 DEBUG SenderThread:2853 [sender.py:send():378] send: telemetry +2024-05-23 12:34:43,498 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: stop_status +2024-05-23 12:34:43,498 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: stop_status +2024-05-23 12:34:43,827 INFO Thread-12 :2853 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/requirements.txt +2024-05-23 12:34:43,827 INFO Thread-12 :2853 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-metadata.json +2024-05-23 12:34:43,828 INFO Thread-12 :2853 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log +2024-05-23 12:34:43,944 INFO wandb-upload_0:2853 [upload_job.py:push():130] Uploaded file /tmp/tmpy2m5btw8wandb/sowjf2hw-wandb-metadata.json +2024-05-23 12:34:45,827 INFO Thread-12 :2853 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log +2024-05-23 12:34:47,563 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:34:52,935 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:34:53,835 INFO Thread-12 :2853 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log +2024-05-23 12:34:54,234 DEBUG SenderThread:2853 [sender.py:send():378] send: exit +2024-05-23 12:34:54,234 INFO SenderThread:2853 [sender.py:send_exit():585] handling exit code: 1 +2024-05-23 12:34:54,234 INFO SenderThread:2853 [sender.py:send_exit():587] handling runtime: 11 +2024-05-23 12:34:54,235 INFO SenderThread:2853 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 12:34:54,236 INFO SenderThread:2853 [sender.py:send_exit():593] send defer +2024-05-23 12:34:54,236 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:54,236 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-23 12:34:54,236 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:54,236 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-23 12:34:54,236 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 1 +2024-05-23 12:34:54,236 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:54,236 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-23 12:34:54,236 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:54,236 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-23 12:34:54,236 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 2 +2024-05-23 12:34:54,236 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:54,236 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-23 12:34:54,236 INFO HandlerThread:2853 [system_monitor.py:finish():203] Stopping system monitor +2024-05-23 12:34:54,237 INFO HandlerThread:2853 [interfaces.py:finish():200] Joined cpu monitor +2024-05-23 12:34:54,237 DEBUG SystemMonitor:2853 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-23 12:34:54,237 DEBUG SystemMonitor:2853 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-23 12:34:54,237 DEBUG SystemMonitor:2853 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-23 12:34:54,240 INFO HandlerThread:2853 [interfaces.py:finish():200] Joined disk monitor +2024-05-23 12:34:54,240 INFO HandlerThread:2853 [interfaces.py:finish():200] Joined memory monitor +2024-05-23 12:34:54,240 INFO HandlerThread:2853 [interfaces.py:finish():200] Joined network monitor +2024-05-23 12:34:54,240 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:54,240 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-23 12:34:54,241 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 3 +2024-05-23 12:34:54,241 DEBUG SenderThread:2853 [sender.py:send():378] send: stats +2024-05-23 12:34:54,242 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:54,242 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-23 12:34:54,242 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:54,242 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-23 12:34:54,242 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 4 +2024-05-23 12:34:54,242 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:54,242 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-23 12:34:54,242 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:54,242 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-23 12:34:54,242 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 5 +2024-05-23 12:34:54,242 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:54,242 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-23 12:34:54,242 DEBUG SenderThread:2853 [sender.py:send():378] send: summary +2024-05-23 12:34:54,243 INFO SenderThread:2853 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-23 12:34:54,243 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:54,243 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-23 12:34:54,243 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 6 +2024-05-23 12:34:54,244 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:54,244 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-23 12:34:54,244 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:54,244 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-23 12:34:54,248 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: status_report +2024-05-23 12:34:54,316 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 7 +2024-05-23 12:34:54,316 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:54,316 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-23 12:34:54,317 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:54,317 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-23 12:34:54,837 INFO Thread-12 :2853 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/config.yaml +2024-05-23 12:34:54,837 INFO Thread-12 :2853 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-summary.json +2024-05-23 12:34:55,234 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:34:55,584 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 8 +2024-05-23 12:34:55,584 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:34:55,584 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:55,585 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-23 12:34:55,585 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:55,585 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-23 12:34:55,585 INFO SenderThread:2853 [job_builder.py:build():432] Attempting to build job artifact +2024-05-23 12:34:55,585 INFO SenderThread:2853 [job_builder.py:_get_source_type():576] no source found +2024-05-23 12:34:55,585 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 9 +2024-05-23 12:34:55,585 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:55,585 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-23 12:34:55,586 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:55,586 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-23 12:34:55,586 INFO SenderThread:2853 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-23 12:34:55,838 INFO SenderThread:2853 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log +2024-05-23 12:34:55,838 INFO SenderThread:2853 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files +2024-05-23 12:34:55,838 INFO SenderThread:2853 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-metadata.json wandb-metadata.json +2024-05-23 12:34:55,839 INFO SenderThread:2853 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log output.log +2024-05-23 12:34:55,839 INFO SenderThread:2853 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/config.yaml config.yaml +2024-05-23 12:34:55,841 INFO SenderThread:2853 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/requirements.txt requirements.txt +2024-05-23 12:34:55,843 INFO SenderThread:2853 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-summary.json wandb-summary.json +2024-05-23 12:34:55,845 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 10 +2024-05-23 12:34:55,846 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:55,846 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-23 12:34:55,846 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:55,848 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-23 12:34:55,848 INFO SenderThread:2853 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 12:34:56,108 INFO wandb-upload_0:2853 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/output.log +2024-05-23 12:34:56,234 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:34:56,235 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:34:56,442 INFO wandb-upload_2:2853 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/requirements.txt +2024-05-23 12:34:56,467 INFO wandb-upload_1:2853 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/config.yaml +2024-05-23 12:34:56,480 INFO wandb-upload_3:2853 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/files/wandb-summary.json +2024-05-23 12:34:56,680 INFO Thread-11 (_thread_body):2853 [sender.py:transition_state():613] send defer: 11 +2024-05-23 12:34:56,680 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:56,680 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-23 12:34:56,680 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:56,680 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-23 12:34:56,680 INFO SenderThread:2853 [file_pusher.py:join():175] waiting for file pusher +2024-05-23 12:34:56,680 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 12 +2024-05-23 12:34:56,681 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:56,681 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-23 12:34:56,681 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:56,681 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-23 12:34:56,681 INFO SenderThread:2853 [file_stream.py:finish():601] file stream finish called +2024-05-23 12:34:56,755 INFO SenderThread:2853 [file_stream.py:finish():605] file stream finish is done +2024-05-23 12:34:56,755 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 13 +2024-05-23 12:34:56,755 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:56,755 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-23 12:34:56,755 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:56,755 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-23 12:34:56,755 INFO SenderThread:2853 [sender.py:transition_state():613] send defer: 14 +2024-05-23 12:34:56,755 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: defer +2024-05-23 12:34:56,755 INFO HandlerThread:2853 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-23 12:34:56,755 DEBUG SenderThread:2853 [sender.py:send():378] send: final +2024-05-23 12:34:56,756 DEBUG SenderThread:2853 [sender.py:send():378] send: footer +2024-05-23 12:34:56,756 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: defer +2024-05-23 12:34:56,756 INFO SenderThread:2853 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-23 12:34:56,756 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:34:56,756 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-23 12:34:56,756 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: server_info +2024-05-23 12:34:56,756 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: get_summary +2024-05-23 12:34:56,757 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-23 12:34:56,757 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-23 12:34:56,757 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:34:56,757 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: poll_exit +2024-05-23 12:34:56,757 DEBUG SenderThread:2853 [sender.py:send_request():405] send_request: server_info +2024-05-23 12:34:56,811 INFO MainThread:2853 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-23 12:34:56,811 INFO MainThread:2853 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-23 12:34:56,811 INFO MainThread:2853 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-23 12:34:56,811 DEBUG HandlerThread:2853 [handler.py:handle_request():158] handle_request: shutdown +2024-05-23 12:34:56,811 INFO HandlerThread:2853 [handler.py:finish():882] shutting down handler +2024-05-23 12:34:57,757 INFO WriterThread:2853 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/run-rpj1bedw.wandb +2024-05-23 12:34:57,811 INFO SenderThread:2853 [sender.py:finish():1545] shutting down sender +2024-05-23 12:34:57,811 INFO SenderThread:2853 [file_pusher.py:finish():169] shutting down file pusher +2024-05-23 12:34:57,811 INFO SenderThread:2853 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/logs/debug.log b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..8a2a28ca58d06b9d11a1952b3c280fae6f2165b3 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_setup.py:_flush():76] Configure stats pid to 2698 +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-23 12:34:42,512 WARNING MainThread:2698 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/logs/debug.log +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/logs/debug-internal.log +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_init.py:init():560] calling init triggers +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_init.py:init():610] starting backend +2024-05-23 12:34:42,512 INFO MainThread:2698 [wandb_init.py:init():614] setting up manager +2024-05-23 12:34:42,515 INFO MainThread:2698 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-23 12:34:42,516 INFO MainThread:2698 [wandb_init.py:init():622] backend started and connected +2024-05-23 12:34:42,520 INFO MainThread:2698 [wandb_init.py:init():711] updated telemetry +2024-05-23 12:34:42,528 INFO MainThread:2698 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-23 12:34:42,828 INFO MainThread:2698 [wandb_run.py:_on_init():2396] communicating current version +2024-05-23 12:34:43,148 INFO MainThread:2698 [wandb_run.py:_on_init():2405] got version response +2024-05-23 12:34:43,148 INFO MainThread:2698 [wandb_init.py:init():795] starting run threads in backend +2024-05-23 12:34:43,454 INFO MainThread:2698 [wandb_run.py:_console_start():2374] atexit reg +2024-05-23 12:34:43,454 INFO MainThread:2698 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-23 12:34:43,454 INFO MainThread:2698 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-23 12:34:43,454 INFO MainThread:2698 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-23 12:34:43,456 INFO MainThread:2698 [wandb_init.py:init():838] run started, returning control to user process +2024-05-23 12:34:57,812 WARNING MsgRouterThr:2698 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/run-rpj1bedw.wandb b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/run-rpj1bedw.wandb new file mode 100644 index 0000000000000000000000000000000000000000..2a6a480f78394a2e88d0831bd5cfe4a6e260be92 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240523_123442-rpj1bedw/run-rpj1bedw.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..6d6fe49051ec02160486b3defb7025cc33dc82b8 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log @@ -0,0 +1,32 @@ + +2024-05-30:08:23:47,282 INFO [__main__.py:251] Verbosity set to INFO +2024-05-30:08:23:56,541 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'boolq', 'copa', 'mrpc', 'piqa', 'sst2', 'winogrande'] +2024-05-30:08:23:56,542 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-05-30:08:23:56,543 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step10000', 'tokenizer': '/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step10000'} +2024-05-30:08:23:58,832 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Traceback (most recent call last): + File "/usr/lib/python3.10/runpy.py", line 196, in _run_module_as_main + return _run_code(code, main_globals, None, + File "/usr/lib/python3.10/runpy.py", line 86, in _run_code + exec(code, run_globals) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 417, in + cli_evaluate() + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/__main__.py", line 341, in cli_evaluate + results = evaluator.simple_evaluate( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/utils.py", line 288, in _wrapper + return fn(*args, **kwargs) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/evaluator.py", line 180, in simple_evaluate + lm = lm_eval.api.registry.get_model(model).create_from_arg_string( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/api/model.py", line 134, in create_from_arg_string + return cls(**args, **args2) + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 237, in __init__ + self._create_tokenizer( + File "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/lm_eval/models/huggingface.py", line 630, in _create_tokenizer + self.tokenizer = transformers.AutoTokenizer.from_pretrained( + File "/usr/local/lib/python3.10/dist-packages/transformers/models/auto/tokenization_auto.py", line 805, in from_pretrained + return tokenizer_class_fast.from_pretrained(pretrained_model_name_or_path, *inputs, **kwargs) + File "/usr/local/lib/python3.10/dist-packages/transformers/tokenization_utils_base.py", line 2012, in from_pretrained + raise EnvironmentError( +OSError: Can't load tokenizer for '/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step10000'. If you were trying to load it from 'https://huggingface.co/models', make sure you don't have a local directory with the same name. Otherwise, make sure '/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step10000' is the correct path to a directory containing all relevant files for a LlamaTokenizerFast tokenizer. \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/wandb-metadata.json b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..42ec97e7dc326612bddf9883988e6b29e4665e81 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/wandb-metadata.json @@ -0,0 +1,850 @@ +{ + "os": "Linux-5.15.0-92-generic-x86_64-with-glibc2.35", + "python": "3.10.12", + "heartbeatAt": "2024-05-30T08:23:47.066028", + "startedAt": "2024-05-30T08:23:46.530676", + "docker": null, + "cuda": null, + "args": [ + "--model", + "hf", + "--model_args", + "pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step10000,tokenizer=/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step10000", + "--tasks", + "winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq", + "--batch_size", + "auto", + "--wandb_args", + "project=english-eval,group=exp2,name=global_step10000" + ], + "state": "running", + "program": "-m lm_eval.__main__", + "codePathLocal": null, + "git": { + "remote": "https://github.com/EleutherAI/lm-evaluation-harness", + "commit": null + }, + "email": null, + "root": "/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness", + "host": "peacock-evaluation-debug-5-worker-0", + "username": "root", + "executable": "/usr/bin/python3", + "cpu_count": 80, + "cpu_count_logical": 160, + "cpu_freq": { + "current": 2327.22031875, + "min": 800.0, + "max": 3400.0 + }, + "cpu_freq_per_core": [ + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 3400.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + }, + { + "current": 2300.0, + "min": 800.0, + "max": 3400.0 + } + ], + "disk": { + "/": { + "total": 877.6341285705566, + "used": 211.90747451782227 + } + }, + "memory": { + "total": 1007.4379997253418 + } +} diff --git a/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/logs/debug-internal.log b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..1db46eccd6a14e775fbc0c062f777087de9e3907 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/logs/debug-internal.log @@ -0,0 +1,196 @@ +2024-05-30 08:23:46,552 INFO StreamThr :903 [internal.py:wandb_internal():85] W&B internal server running at pid: 903, started at: 2024-05-30 08:23:46.550229 +2024-05-30 08:23:46,556 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status +2024-05-30 08:23:46,556 INFO WriterThread:903 [datastore.py:open_for_write():87] open: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/run-v7juco5r.wandb +2024-05-30 08:23:46,559 DEBUG SenderThread:903 [sender.py:send():378] send: header +2024-05-30 08:23:46,563 DEBUG SenderThread:903 [sender.py:send():378] send: run +2024-05-30 08:23:46,872 INFO SenderThread:903 [dir_watcher.py:__init__():211] watching files in: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files +2024-05-30 08:23:46,872 INFO SenderThread:903 [sender.py:_start_run_threads():1123] run started: v7juco5r with start time 1717057426.550073 +2024-05-30 08:23:46,876 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: check_version +2024-05-30 08:23:46,876 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: check_version +2024-05-30 08:23:46,992 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: run_start +2024-05-30 08:23:46,994 DEBUG HandlerThread:903 [system_info.py:__init__():26] System info init +2024-05-30 08:23:46,995 DEBUG HandlerThread:903 [system_info.py:__init__():41] System info init done +2024-05-30 08:23:46,995 INFO HandlerThread:903 [system_monitor.py:start():194] Starting system monitor +2024-05-30 08:23:46,995 INFO SystemMonitor:903 [system_monitor.py:_start():158] Starting system asset monitoring threads +2024-05-30 08:23:46,995 INFO HandlerThread:903 [system_monitor.py:probe():214] Collecting system info +2024-05-30 08:23:47,002 INFO SystemMonitor:903 [interfaces.py:start():188] Started cpu monitoring +2024-05-30 08:23:47,002 INFO SystemMonitor:903 [interfaces.py:start():188] Started disk monitoring +2024-05-30 08:23:47,008 INFO SystemMonitor:903 [interfaces.py:start():188] Started memory monitoring +2024-05-30 08:23:47,008 INFO SystemMonitor:903 [interfaces.py:start():188] Started network monitoring +2024-05-30 08:23:47,065 DEBUG HandlerThread:903 [system_info.py:probe():150] Probing system +2024-05-30 08:23:47,069 DEBUG HandlerThread:903 [system_info.py:_probe_git():135] Probing git +2024-05-30 08:23:47,079 ERROR HandlerThread:903 [gitlib.py:root():92] git root error: Cmd('git') failed due to: exit code(128) + cmdline: git rev-parse --show-toplevel + stderr: 'fatal: detected dubious ownership in repository at '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +To add an exception for this directory, call: + + git config --global --add safe.directory /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness' +2024-05-30 08:23:47,079 DEBUG HandlerThread:903 [system_info.py:_probe_git():143] Probing git done +2024-05-30 08:23:47,079 DEBUG HandlerThread:903 [system_info.py:probe():198] Probing system done +2024-05-30 08:23:47,079 DEBUG HandlerThread:903 [system_monitor.py:probe():223] {'os': 'Linux-5.15.0-92-generic-x86_64-with-glibc2.35', 'python': '3.10.12', 'heartbeatAt': '2024-05-30T08:23:47.066028', 'startedAt': '2024-05-30T08:23:46.530676', 'docker': None, 'cuda': None, 'args': ('--model', 'hf', '--model_args', 'pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step10000,tokenizer=/mnt/weka/peacock/experiments/llama/eval/checkpoint-english/llamav2-3b/hf/global_step10000', '--tasks', 'winogrande,sst2,mrpc,arc_easy,copa,piqa,boolq', '--batch_size', 'auto', '--wandb_args', 'project=english-eval,group=exp2,name=global_step10000'), 'state': 'running', 'program': '-m lm_eval.__main__', 'codePathLocal': None, 'git': {'remote': 'https://github.com/EleutherAI/lm-evaluation-harness', 'commit': None}, 'email': None, 'root': '/mnt/weka/peacock/idc/cronscript/lm-evaluation-harness', 'host': 'peacock-evaluation-debug-5-worker-0', 'username': 'root', 'executable': '/usr/bin/python3', 'cpu_count': 80, 'cpu_count_logical': 160, 'cpu_freq': {'current': 2327.22031875, 'min': 800.0, 'max': 3400.0}, 'cpu_freq_per_core': [{'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 3400.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}, {'current': 2300.0, 'min': 800.0, 'max': 3400.0}], 'disk': {'/': {'total': 877.6341285705566, 'used': 211.90747451782227}}, 'memory': {'total': 1007.4379997253418}} +2024-05-30 08:23:47,079 INFO HandlerThread:903 [system_monitor.py:probe():224] Finished collecting system info +2024-05-30 08:23:47,079 INFO HandlerThread:903 [system_monitor.py:probe():227] Publishing system info +2024-05-30 08:23:47,083 INFO HandlerThread:903 [system_monitor.py:probe():229] Finished publishing system info +2024-05-30 08:23:47,090 DEBUG SenderThread:903 [sender.py:send():378] send: files +2024-05-30 08:23:47,090 INFO SenderThread:903 [sender.py:_save_file():1389] saving file wandb-metadata.json with policy now +2024-05-30 08:23:47,274 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: python_packages +2024-05-30 08:23:47,274 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: python_packages +2024-05-30 08:23:47,275 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 08:23:47,277 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: stop_status +2024-05-30 08:23:47,437 DEBUG SenderThread:903 [sender.py:send():378] send: telemetry +2024-05-30 08:23:47,709 INFO wandb-upload_0:903 [upload_job.py:push():130] Uploaded file /tmp/tmpj1_qaha_wandb/11oukotf-wandb-metadata.json +2024-05-30 08:23:47,875 INFO Thread-12 :903 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log +2024-05-30 08:23:47,876 INFO Thread-12 :903 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/requirements.txt +2024-05-30 08:23:47,876 INFO Thread-12 :903 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/wandb-metadata.json +2024-05-30 08:23:49,875 INFO Thread-12 :903 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log +2024-05-30 08:23:52,440 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 08:23:57,544 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 08:23:57,882 INFO Thread-12 :903 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log +2024-05-30 08:23:59,887 INFO Thread-12 :903 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log +2024-05-30 08:24:02,275 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 08:24:02,276 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: stop_status +2024-05-30 08:24:03,376 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 08:24:08,377 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 08:24:13,377 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 08:24:17,275 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: stop_status +2024-05-30 08:24:17,276 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: stop_status +2024-05-30 08:24:18,442 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 08:24:18,928 INFO Thread-12 :903 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/config.yaml +2024-05-30 08:24:24,230 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 08:24:24,274 INFO Thread-12 :903 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log +2024-05-30 08:24:25,128 DEBUG SenderThread:903 [sender.py:send():378] send: exit +2024-05-30 08:24:25,128 INFO SenderThread:903 [sender.py:send_exit():585] handling exit code: 1 +2024-05-30 08:24:25,128 INFO SenderThread:903 [sender.py:send_exit():587] handling runtime: 38 +2024-05-30 08:24:25,130 INFO SenderThread:903 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-30 08:24:25,130 INFO SenderThread:903 [sender.py:send_exit():593] send defer +2024-05-30 08:24:25,130 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:25,130 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 0 +2024-05-30 08:24:25,130 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:25,130 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 0 +2024-05-30 08:24:25,130 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 1 +2024-05-30 08:24:25,130 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:25,130 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 1 +2024-05-30 08:24:25,130 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:25,131 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 1 +2024-05-30 08:24:25,131 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 2 +2024-05-30 08:24:25,131 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:25,131 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 2 +2024-05-30 08:24:25,131 INFO HandlerThread:903 [system_monitor.py:finish():203] Stopping system monitor +2024-05-30 08:24:25,131 DEBUG SystemMonitor:903 [system_monitor.py:_start():172] Starting system metrics aggregation loop +2024-05-30 08:24:25,131 DEBUG SystemMonitor:903 [system_monitor.py:_start():179] Finished system metrics aggregation loop +2024-05-30 08:24:25,131 DEBUG SystemMonitor:903 [system_monitor.py:_start():183] Publishing last batch of metrics +2024-05-30 08:24:25,134 INFO HandlerThread:903 [interfaces.py:finish():200] Joined cpu monitor +2024-05-30 08:24:25,134 INFO HandlerThread:903 [interfaces.py:finish():200] Joined disk monitor +2024-05-30 08:24:25,134 INFO HandlerThread:903 [interfaces.py:finish():200] Joined memory monitor +2024-05-30 08:24:25,134 INFO HandlerThread:903 [interfaces.py:finish():200] Joined network monitor +2024-05-30 08:24:25,135 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:25,135 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 2 +2024-05-30 08:24:25,135 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 3 +2024-05-30 08:24:25,135 DEBUG SenderThread:903 [sender.py:send():378] send: stats +2024-05-30 08:24:25,136 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:25,136 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 3 +2024-05-30 08:24:25,136 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:25,136 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 3 +2024-05-30 08:24:25,136 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 4 +2024-05-30 08:24:25,136 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:25,136 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 4 +2024-05-30 08:24:25,136 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:25,136 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 4 +2024-05-30 08:24:25,136 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 5 +2024-05-30 08:24:25,136 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:25,136 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 5 +2024-05-30 08:24:25,137 DEBUG SenderThread:903 [sender.py:send():378] send: summary +2024-05-30 08:24:25,137 INFO SenderThread:903 [sender.py:_save_file():1389] saving file wandb-summary.json with policy end +2024-05-30 08:24:25,137 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:25,138 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 5 +2024-05-30 08:24:25,138 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 6 +2024-05-30 08:24:25,138 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:25,138 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 6 +2024-05-30 08:24:25,138 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:25,138 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 6 +2024-05-30 08:24:25,138 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 7 +2024-05-30 08:24:25,138 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: status_report +2024-05-30 08:24:25,138 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:25,138 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 7 +2024-05-30 08:24:25,138 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:25,138 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 7 +2024-05-30 08:24:25,276 INFO Thread-12 :903 [dir_watcher.py:_on_file_created():271] file/dir created: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/wandb-summary.json +2024-05-30 08:24:26,128 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 08:24:27,486 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 8 +2024-05-30 08:24:27,486 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 08:24:27,486 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:27,486 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 8 +2024-05-30 08:24:27,487 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:27,487 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 8 +2024-05-30 08:24:27,487 INFO SenderThread:903 [job_builder.py:build():432] Attempting to build job artifact +2024-05-30 08:24:27,487 INFO SenderThread:903 [job_builder.py:_get_source_type():576] no source found +2024-05-30 08:24:27,488 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 9 +2024-05-30 08:24:27,488 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:27,488 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 9 +2024-05-30 08:24:27,488 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:27,488 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 9 +2024-05-30 08:24:27,488 INFO SenderThread:903 [dir_watcher.py:finish():358] shutting down directory watcher +2024-05-30 08:24:28,129 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 08:24:28,277 INFO SenderThread:903 [dir_watcher.py:_on_file_modified():288] file/dir modified: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log +2024-05-30 08:24:28,277 INFO SenderThread:903 [dir_watcher.py:finish():388] scan: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files +2024-05-30 08:24:28,278 INFO SenderThread:903 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/wandb-metadata.json wandb-metadata.json +2024-05-30 08:24:28,278 INFO SenderThread:903 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/config.yaml config.yaml +2024-05-30 08:24:28,278 INFO SenderThread:903 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/requirements.txt requirements.txt +2024-05-30 08:24:28,280 INFO SenderThread:903 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/wandb-summary.json wandb-summary.json +2024-05-30 08:24:28,281 INFO SenderThread:903 [dir_watcher.py:finish():402] scan save: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log output.log +2024-05-30 08:24:28,281 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 10 +2024-05-30 08:24:28,281 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 08:24:28,281 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:28,281 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 10 +2024-05-30 08:24:28,281 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:28,281 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 10 +2024-05-30 08:24:28,281 INFO SenderThread:903 [file_pusher.py:finish():169] shutting down file pusher +2024-05-30 08:24:28,685 INFO wandb-upload_0:903 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/config.yaml +2024-05-30 08:24:28,831 INFO wandb-upload_1:903 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/requirements.txt +2024-05-30 08:24:28,907 INFO wandb-upload_2:903 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/wandb-summary.json +2024-05-30 08:24:28,922 INFO wandb-upload_3:903 [upload_job.py:push():130] Uploaded file /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/files/output.log +2024-05-30 08:24:29,123 INFO Thread-11 (_thread_body):903 [sender.py:transition_state():613] send defer: 11 +2024-05-30 08:24:29,123 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:29,123 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 11 +2024-05-30 08:24:29,123 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:29,123 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 11 +2024-05-30 08:24:29,123 INFO SenderThread:903 [file_pusher.py:join():175] waiting for file pusher +2024-05-30 08:24:29,123 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 12 +2024-05-30 08:24:29,124 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:29,124 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 12 +2024-05-30 08:24:29,124 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:29,124 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 12 +2024-05-30 08:24:29,124 INFO SenderThread:903 [file_stream.py:finish():601] file stream finish called +2024-05-30 08:24:29,129 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 08:24:29,357 INFO SenderThread:903 [file_stream.py:finish():605] file stream finish is done +2024-05-30 08:24:29,357 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 13 +2024-05-30 08:24:29,357 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 08:24:29,357 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:29,357 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 13 +2024-05-30 08:24:29,357 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:29,357 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 13 +2024-05-30 08:24:29,357 INFO SenderThread:903 [sender.py:transition_state():613] send defer: 14 +2024-05-30 08:24:29,357 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: defer +2024-05-30 08:24:29,357 INFO HandlerThread:903 [handler.py:handle_request_defer():184] handle defer: 14 +2024-05-30 08:24:29,358 DEBUG SenderThread:903 [sender.py:send():378] send: final +2024-05-30 08:24:29,358 DEBUG SenderThread:903 [sender.py:send():378] send: footer +2024-05-30 08:24:29,358 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: defer +2024-05-30 08:24:29,358 INFO SenderThread:903 [sender.py:send_request_defer():609] handle sender defer: 14 +2024-05-30 08:24:29,358 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 08:24:29,358 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 08:24:29,360 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: poll_exit +2024-05-30 08:24:29,360 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: server_info +2024-05-30 08:24:29,360 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: get_summary +2024-05-30 08:24:29,360 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: sampled_history +2024-05-30 08:24:29,360 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: internal_messages +2024-05-30 08:24:29,360 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: poll_exit +2024-05-30 08:24:29,360 DEBUG SenderThread:903 [sender.py:send_request():405] send_request: server_info +2024-05-30 08:24:29,413 INFO MainThread:903 [wandb_run.py:_footer_history_summary_info():3994] rendering history +2024-05-30 08:24:29,413 INFO MainThread:903 [wandb_run.py:_footer_history_summary_info():4026] rendering summary +2024-05-30 08:24:29,413 INFO MainThread:903 [wandb_run.py:_footer_sync_info():3953] logging synced files +2024-05-30 08:24:29,414 DEBUG HandlerThread:903 [handler.py:handle_request():158] handle_request: shutdown +2024-05-30 08:24:29,414 INFO HandlerThread:903 [handler.py:finish():882] shutting down handler +2024-05-30 08:24:30,360 INFO WriterThread:903 [datastore.py:close():296] close: /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/run-v7juco5r.wandb +2024-05-30 08:24:30,413 INFO SenderThread:903 [sender.py:finish():1545] shutting down sender +2024-05-30 08:24:30,413 INFO SenderThread:903 [file_pusher.py:finish():169] shutting down file pusher +2024-05-30 08:24:30,413 INFO SenderThread:903 [file_pusher.py:join():175] waiting for file pusher diff --git a/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/logs/debug.log b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..2a67b19f4a46fd99b9f6aef0ca2168d02ae51d16 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/logs/debug.log @@ -0,0 +1,29 @@ +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_setup.py:_flush():76] Current SDK version is 0.17.0 +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_setup.py:_flush():76] Configure stats pid to 747 +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_setup.py:_flush():76] Loading settings from /root/.config/wandb/settings +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_setup.py:_flush():76] Loading settings from /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/settings +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_setup.py:_flush():76] Loading settings from environment variables: {} +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_setup.py:_flush():76] Applying setup settings: {'_disable_service': False} +2024-05-30 08:23:46,544 WARNING MainThread:747 [wandb_setup.py:_flush():76] Could not find program at -m lm_eval.__main__ +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_setup.py:_flush():76] Inferring run settings from compute environment: {'program_relpath': None, 'program': '-m lm_eval.__main__'} +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_setup.py:_flush():76] Applying login settings: {} +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_init.py:_log_setup():520] Logging user logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/logs/debug.log +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_init.py:_log_setup():521] Logging internal logs to /mnt/weka/peacock/idc/cronscript/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/logs/debug-internal.log +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_init.py:init():560] calling init triggers +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_init.py:init():567] wandb.init called with sweep_config: {} +config: {} +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_init.py:init():610] starting backend +2024-05-30 08:23:46,544 INFO MainThread:747 [wandb_init.py:init():614] setting up manager +2024-05-30 08:23:46,548 INFO MainThread:747 [backend.py:_multiprocessing_setup():105] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2024-05-30 08:23:46,549 INFO MainThread:747 [wandb_init.py:init():622] backend started and connected +2024-05-30 08:23:46,553 INFO MainThread:747 [wandb_init.py:init():711] updated telemetry +2024-05-30 08:23:46,562 INFO MainThread:747 [wandb_init.py:init():744] communicating run to backend with 90.0 second timeout +2024-05-30 08:23:46,875 INFO MainThread:747 [wandb_run.py:_on_init():2396] communicating current version +2024-05-30 08:23:46,986 INFO MainThread:747 [wandb_run.py:_on_init():2405] got version response +2024-05-30 08:23:46,986 INFO MainThread:747 [wandb_init.py:init():795] starting run threads in backend +2024-05-30 08:23:47,275 INFO MainThread:747 [wandb_run.py:_console_start():2374] atexit reg +2024-05-30 08:23:47,276 INFO MainThread:747 [wandb_run.py:_redirect():2229] redirect: wrap_raw +2024-05-30 08:23:47,276 INFO MainThread:747 [wandb_run.py:_redirect():2294] Wrapping output streams. +2024-05-30 08:23:47,276 INFO MainThread:747 [wandb_run.py:_redirect():2319] Redirects installed. +2024-05-30 08:23:47,279 INFO MainThread:747 [wandb_init.py:init():838] run started, returning control to user process +2024-05-30 08:24:30,414 WARNING MsgRouterThr:747 [router.py:message_loop():77] message_loop has been closed diff --git a/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/run-v7juco5r.wandb b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/run-v7juco5r.wandb new file mode 100644 index 0000000000000000000000000000000000000000..b3a45ed4dbcc765f92de9e77bbfce1b8488f81f1 Binary files /dev/null and b/lm-evaluation-harness/wandb/run-20240530_082346-v7juco5r/run-v7juco5r.wandb differ diff --git a/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/config.yaml b/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..54e7ef981baa242a387c8519ac614c2b636372ce --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/config.yaml @@ -0,0 +1,375 @@ +wandb_version: 1 + +_wandb: + desc: null + value: + python_version: 3.10.12 + cli_version: 0.17.0 + framework: huggingface + huggingface_version: 4.36.2 + is_jupyter_run: false + is_kaggle_kernel: false + start_time: 1717585238 + t: + 1: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 2: + - 1 + - 5 + - 11 + - 49 + - 51 + - 53 + - 55 + - 71 + - 98 + - 100 + 3: + - 2 + - 13 + - 23 + - 62 + 4: 3.10.12 + 5: 0.17.0 + 6: 4.36.2 + 8: + - 5 + 13: linux-x86_64 +task_configs: + desc: null + value: + arc_easy: + task: arc_easy + group: + - ai2_arc + dataset_path: allenai/ai2_arc + dataset_name: ARC-Easy + training_split: train + validation_split: validation + test_split: test + doc_to_text: 'Question: {{question}} + + Answer:' + doc_to_target: '{{choices.label.index(answerKey)}}' + doc_to_choice: '{{choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{question}} + + Answer:' + metadata: + version: 1.0 + boolq: + task: boolq + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: boolq + training_split: train + validation_split: validation + doc_to_text: '{{passage}} + + Question: {{question}}? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: passage + metadata: + version: 2.0 + copa: + task: copa + group: + - super-glue-lm-eval-v1 + dataset_path: super_glue + dataset_name: copa + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n # Drop the period\n connector =\ + \ {\n \"cause\": \"because\",\n \"effect\": \"therefore\",\n\ + \ }[doc[\"question\"]]\n return doc[\"premise\"].strip()[:-1] + f\"\ + \ {connector}\"\n" + doc_to_target: "def doc_to_target(doc):\n correct_choice = doc[\"choice1\"\ + ] if doc[\"label\"] == 0 else doc[\"choice2\"]\n # Connect the sentences\n\ + \ return \" \" + convert_choice(correct_choice)\n" + doc_to_choice: "def doc_to_choice(doc):\n return [\" \" + convert_choice(doc[\"\ + choice1\"]), \" \" + convert_choice(doc[\"choice2\"])]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + indic_arc_challenge_hi: + task: indic_arc_challenge_hi + group: Cognitive-Lab/Indic-ARC-Challenge + dataset_path: Cognitive-Lab/Indic-ARC-Challenge + dataset_name: hi + test_split: test + doc_to_text: 'Question: {{translated_question}} + + Answer:' + doc_to_target: '{{translated_choices.label.index(answerKey)}}' + doc_to_choice: '{{translated_choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{translated_question}} + + Answer:' + metadata: + version: 1.0 + indic_arc_easy_hi: + task: indic_arc_easy_hi + group: Cognitive-Lab/Indic-ARC-Easy + dataset_path: Cognitive-Lab/Indic-ARC-Easy + dataset_name: hi + test_split: test + doc_to_text: 'Question: {{translated_question}} + + Answer:' + doc_to_target: '{{translated_choices.label.index(answerKey)}}' + doc_to_choice: '{{translated_choices.text}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: 'Question: {{translated_question}} + + Answer:' + metadata: + version: 1.0 + indic_boolq_hi: + task: indic_boolq_hi + group: Cognitive-Lab/Indic-BoolQ + dataset_path: Cognitive-Lab/Indic-BoolQ + dataset_name: hi + validation_split: validation + doc_to_text: 'Passage: {translated_passage} + + Question: {translated_question.strip()} + + Answer:' + doc_to_target: answer + doc_to_choice: + - 'true' + - 'false' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + mrpc: + task: mrpc + group: glue + dataset_path: glue + dataset_name: mrpc + training_split: train + validation_split: validation + doc_to_text: 'Sentence 1: {{sentence1}} + + Sentence 2: {{sentence2}} + + Question: Do both sentences mean the same thing? + + Answer:' + doc_to_target: label + doc_to_choice: + - 'no' + - 'yes' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + - metric: f1 + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + piqa: + task: piqa + dataset_path: piqa + training_split: train + validation_split: validation + doc_to_text: 'Question: {{goal}} + + Answer:' + doc_to_target: label + doc_to_choice: '{{[sol1, sol2]}}' + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + - metric: acc_norm + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: goal + metadata: + version: 1.0 + sst2: + task: sst2 + group: glue + dataset_path: glue + dataset_name: sst2 + training_split: train + validation_split: validation + doc_to_text: '{{sentence}} + + Question: Is this sentence positive or negative? + + Answer:' + doc_to_target: label + doc_to_choice: + - negative + - positive + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + output_type: multiple_choice + repeats: 1 + should_decontaminate: false + metadata: + version: 1.0 + winogrande: + task: winogrande + dataset_path: winogrande + dataset_name: winogrande_xl + training_split: train + validation_split: validation + doc_to_text: "def doc_to_text(doc):\n answer_to_num = {\"1\": 0, \"2\": 1}\n\ + \ return answer_to_num[doc[\"answer\"]]\n" + doc_to_target: "def doc_to_target(doc):\n idx = doc[\"sentence\"].index(\"\ + _\") + 1\n return doc[\"sentence\"][idx:].strip()\n" + doc_to_choice: "def doc_to_choice(doc):\n idx = doc[\"sentence\"].index(\"\ + _\")\n options = [doc[\"option1\"], doc[\"option2\"]]\n return [doc[\"\ + sentence\"][:idx] + opt for opt in options]\n" + description: '' + target_delimiter: ' ' + fewshot_delimiter: ' + + + ' + num_fewshot: 0 + metric_list: + - metric: acc + aggregation: mean + higher_is_better: true + output_type: multiple_choice + repeats: 1 + should_decontaminate: true + doc_to_decontamination_query: sentence + metadata: + version: 1.0 +cli_configs: + desc: null + value: + model: hf + model_args: pretrained=/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn/llamav2-3b/hf/global_step76000,tokenizer=/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer + batch_size: auto + batch_sizes: + - 64 + device: null + use_cache: null + limit: null + bootstrap_iters: 100000 + gen_kwargs: null diff --git a/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/media/table/evaluation/eval_results_1_1b420fa8f5f420ab8fe6.table.json b/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/media/table/evaluation/eval_results_1_1b420fa8f5f420ab8fe6.table.json new file mode 100644 index 0000000000000000000000000000000000000000..5107bfb2c97c17598d1542eb7ffdf535712bbf56 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/media/table/evaluation/eval_results_1_1b420fa8f5f420ab8fe6.table.json @@ -0,0 +1 @@ +{"columns": ["Tasks", "Version", "Filter", "num_fewshot", "Metric", "Value", "Stderr"], "data": [["winogrande", 1.0, "none", 0, "acc", "0.5035516969218626", "0.0141"], ["sst2", 1.0, "none", 0, "acc", "0.5034403669724771", "0.0169"], ["piqa", 1.0, "none", 0, "acc", "0.5244831338411317", "0.0117"], ["piqa", 1.0, "none", 0, "acc_norm", "0.49183895538628947", "0.0117"], ["mrpc", 1.0, "none", 0, "acc", "0.3161764705882353", "0.0230"], ["mrpc", 1.0, "none", 0, "f1", "0.0", "0.0000"], ["indic_boolq_hi", 1.0, "none", 0, "acc", "0.6217125382262997", "0.0085"], ["indic_arc_easy_hi", 1.0, "none", 0, "acc", "0.24579124579124578", "0.0088"], ["indic_arc_challenge_hi", 1.0, "none", 0, "acc", "0.22098976109215018", "0.0121"], ["copa", 1.0, "none", 0, "acc", "0.56", "0.0499"], ["boolq", 2.0, "none", 0, "acc", "0.3779816513761468", "0.0085"], ["arc_easy", 1.0, "none", 0, "acc", "0.257996632996633", "0.0090"], ["arc_easy", 1.0, "none", 0, "acc_norm", "0.2668350168350168", "0.0091"]]} \ No newline at end of file diff --git a/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/output.log b/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..85527056fd114db021ccb9a42492ec843b6722b6 --- /dev/null +++ b/lm-evaluation-harness/wandb/run-20240605_110038-lwgnduuo/files/output.log @@ -0,0 +1,759 @@ + +2024-06-05:11:00:39,430 INFO [__main__.py:251] Verbosity set to INFO +2024-06-05:11:00:48,563 INFO [__main__.py:335] Selected Tasks: ['arc_easy', 'boolq', 'copa', 'indic_arc_challenge_hi', 'indic_arc_easy_hi', 'indic_boolq_hi', 'mrpc', 'piqa', 'sst2', 'winogrande'] +2024-06-05:11:00:48,564 INFO [evaluator.py:131] Setting random seed to 0 | Setting numpy seed to 1234 | Setting torch manual seed to 1234 +2024-06-05:11:00:48,564 INFO [evaluator.py:177] Initializing hf model, with arguments: {'pretrained': '/mnt/weka/peacock/experiments/llama/eval/checkpoint-enhibn/llamav2-3b/hf/global_step76000', 'tokenizer': '/mnt/weka/peacock/tokenization/trained-tokenizer/enhiben_50k_hf/ConvertedTokenizer'} +2024-06-05:11:00:50,878 INFO [huggingface.py:164] Using device 'cuda' +/usr/local/lib/python3.10/dist-packages/habana_frameworks/torch/gpu_migration/torch/cuda/memory.py:36: UserWarning: No need to call empty_cache on HPU. It manages the memory internally in an effcient way. + warnings.warn( +Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained. +2024-06-05:11:01:20,197 WARNING [task.py:763] [Task: boolq] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-05:11:01:20,198 WARNING [task.py:775] [Task: boolq] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for super_glue contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/super_glue +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-06-05:11:01:22,578 WARNING [task.py:763] [Task: copa] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-05:11:01:22,578 WARNING [task.py:775] [Task: copa] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +2024-06-05:11:01:26,992 WARNING [task.py:322] [Task: indic_arc_challenge_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-05:11:01:26,993 WARNING [task.py:322] [Task: indic_arc_challenge_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-05:11:01:29,544 WARNING [task.py:322] [Task: indic_arc_easy_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-05:11:01:29,544 WARNING [task.py:322] [Task: indic_arc_easy_hi] has_training_docs and has_validation_docs are False, using test_docs as fewshot_docs but this is not recommended. +2024-06-05:11:01:32,070 WARNING [task.py:763] [Task: mrpc] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-05:11:01:32,070 WARNING [task.py:775] [Task: mrpc] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +2024-06-05:11:01:32,071 WARNING [task.py:763] [Task: mrpc] metric f1 is defined, but aggregation is not. using default aggregation=f1 +2024-06-05:11:01:32,071 WARNING [task.py:775] [Task: mrpc] metric f1 is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for piqa contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/piqa +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-06-05:11:01:41,200 WARNING [task.py:763] [Task: sst2] metric acc is defined, but aggregation is not. using default aggregation=mean +2024-06-05:11:01:41,201 WARNING [task.py:775] [Task: sst2] metric acc is defined, but higher_is_better is not. using default higher_is_better=True +/usr/local/lib/python3.10/dist-packages/datasets/load.py:1491: FutureWarning: The repository for winogrande contains custom code which must be executed to correctly load the dataset. You can inspect the repository content at https://hf.co/datasets/winogrande +You can avoid this message in future by passing the argument `trust_remote_code=True`. +Passing `trust_remote_code=True` will be mandatory to load this dataset from the next major release of `datasets`. + warnings.warn( +2024-06-05:11:01:50,773 INFO [task.py:395] Building contexts for winogrande on rank 0... +100%|██████████| 1267/1267 [00:00<00:00, 68736.85it/s] +2024-06-05:11:01:50,859 INFO [task.py:395] Building contexts for sst2 on rank 0... +100%|██████████| 872/872 [00:00<00:00, 2568.93it/s] +2024-06-05:11:01:51,227 INFO [task.py:395] Building contexts for piqa on rank 0... +100%|██████████| 1838/1838 [00:01<00:00, 1090.70it/s] +2024-06-05:11:01:52,990 INFO [task.py:395] Building contexts for mrpc on rank 0... +100%|██████████| 408/408 [00:00<00:00, 1865.39it/s] +2024-06-05:11:01:53,227 INFO [task.py:395] Building contexts for indic_boolq_hi on rank 0... +100%|██████████| 3270/3270 [00:00<00:00, 3606.07it/s] +2024-06-05:11:01:54,310 INFO [task.py:395] Building contexts for indic_arc_easy_hi on rank 0... +100%|██████████| 2376/2376 [00:02<00:00, 1134.44it/s] +2024-06-05:11:01:56,806 INFO [task.py:395] Building contexts for indic_arc_challenge_hi on rank 0... +100%|██████████| 1172/1172 [00:01<00:00, 1127.47it/s] +2024-06-05:11:01:57,968 INFO [task.py:395] Building contexts for copa on rank 0... +100%|██████████| 100/100 [00:00<00:00, 60760.60it/s] +2024-06-05:11:01:57,977 INFO [task.py:395] Building contexts for boolq on rank 0... +100%|██████████| 3270/3270 [00:01<00:00, 1992.57it/s] +2024-06-05:11:01:59,750 INFO [task.py:395] Building contexts for arc_easy on rank 0... +100%|██████████| 2376/2376 [00:02<00:00, 1052.84it/s] +2024-06-05:11:02:02,153 INFO [evaluator.py:379] Running loglikelihood requests +Token indices sequence length is longer than the specified maximum sequence length for this model (1333 > 1024). Running this sequence through the model will result in indexing errors +Running loglikelihood requests: 0%| | 0/45739 [00:00 1024). Running this sequence through the model will result in indexing errors +Running loglikelihood requests: 0%| | 0/45739 [00:00